[ 473.978223] env[63372]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=63372) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 473.978565] env[63372]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=63372) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 473.978681] env[63372]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=63372) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 473.978966] env[63372]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 474.074940] env[63372]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=63372) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 474.085197] env[63372]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=63372) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 474.685231] env[63372]: INFO nova.virt.driver [None req-0d619f09-09b8-42f2-9d19-126708814b69 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 474.754072] env[63372]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 474.754482] env[63372]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 474.754700] env[63372]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=63372) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 477.829603] env[63372]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-98617718-fb11-419b-9bc0-899c0735c034 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.845377] env[63372]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=63372) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 477.845513] env[63372]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-0ca6611a-a9e0-4afe-85e3-83364a351f77 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.871470] env[63372]: INFO oslo_vmware.api [-] Successfully established new session; session ID is def72. [ 477.871610] env[63372]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.117s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 477.872134] env[63372]: INFO nova.virt.vmwareapi.driver [None req-0d619f09-09b8-42f2-9d19-126708814b69 None None] VMware vCenter version: 7.0.3 [ 477.875530] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11144d97-2ccf-4c51-87c2-1be1e2b16357 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.892844] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9aab7c6-d074-4231-a2e1-ded9b4cc1c7e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.898631] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a8b2fd-0b12-429d-a747-620763675047 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.904923] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96632375-ea38-4514-8bdf-2ac968cddbb4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.917622] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f490e6c-5666-40f0-b5d2-8c16dbc5bd33 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.923303] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-025c8684-334a-4367-9ae7-b07482802049 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.952765] env[63372]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-5e92f379-d3b6-4302-acb5-8ce9eceeabcd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.957722] env[63372]: DEBUG nova.virt.vmwareapi.driver [None req-0d619f09-09b8-42f2-9d19-126708814b69 None None] Extension org.openstack.compute already exists. {{(pid=63372) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 477.960362] env[63372]: INFO nova.compute.provider_config [None req-0d619f09-09b8-42f2-9d19-126708814b69 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 478.464072] env[63372]: DEBUG nova.context [None req-0d619f09-09b8-42f2-9d19-126708814b69 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),0b265c72-67f2-4026-92b2-60ffa767eec2(cell1) {{(pid=63372) load_cells /opt/stack/nova/nova/context.py:464}} [ 478.466244] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 478.466508] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 478.467197] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 478.467642] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Acquiring lock "0b265c72-67f2-4026-92b2-60ffa767eec2" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 478.467831] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Lock "0b265c72-67f2-4026-92b2-60ffa767eec2" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 478.468845] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Lock "0b265c72-67f2-4026-92b2-60ffa767eec2" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 478.489741] env[63372]: INFO dbcounter [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Registered counter for database nova_cell0 [ 478.497842] env[63372]: INFO dbcounter [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Registered counter for database nova_cell1 [ 478.501207] env[63372]: DEBUG oslo_db.sqlalchemy.engines [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63372) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 478.501557] env[63372]: DEBUG oslo_db.sqlalchemy.engines [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63372) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 478.506335] env[63372]: ERROR nova.db.main.api [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 478.506335] env[63372]: result = function(*args, **kwargs) [ 478.506335] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 478.506335] env[63372]: return func(*args, **kwargs) [ 478.506335] env[63372]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 478.506335] env[63372]: result = fn(*args, **kwargs) [ 478.506335] env[63372]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 478.506335] env[63372]: return f(*args, **kwargs) [ 478.506335] env[63372]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 478.506335] env[63372]: return db.service_get_minimum_version(context, binaries) [ 478.506335] env[63372]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 478.506335] env[63372]: _check_db_access() [ 478.506335] env[63372]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 478.506335] env[63372]: stacktrace = ''.join(traceback.format_stack()) [ 478.506335] env[63372]: [ 478.507312] env[63372]: ERROR nova.db.main.api [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 478.507312] env[63372]: result = function(*args, **kwargs) [ 478.507312] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 478.507312] env[63372]: return func(*args, **kwargs) [ 478.507312] env[63372]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 478.507312] env[63372]: result = fn(*args, **kwargs) [ 478.507312] env[63372]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 478.507312] env[63372]: return f(*args, **kwargs) [ 478.507312] env[63372]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 478.507312] env[63372]: return db.service_get_minimum_version(context, binaries) [ 478.507312] env[63372]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 478.507312] env[63372]: _check_db_access() [ 478.507312] env[63372]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 478.507312] env[63372]: stacktrace = ''.join(traceback.format_stack()) [ 478.507312] env[63372]: [ 478.507708] env[63372]: WARNING nova.objects.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 478.507828] env[63372]: WARNING nova.objects.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Failed to get minimum service version for cell 0b265c72-67f2-4026-92b2-60ffa767eec2 [ 478.508255] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Acquiring lock "singleton_lock" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 478.508417] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Acquired lock "singleton_lock" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 478.508659] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Releasing lock "singleton_lock" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 478.508973] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Full set of CONF: {{(pid=63372) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 478.509132] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ******************************************************************************** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 478.509262] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Configuration options gathered from: {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 478.509402] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 478.509592] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 478.509720] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ================================================================================ {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 478.509926] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] allow_resize_to_same_host = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.510110] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] arq_binding_timeout = 300 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.510246] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] backdoor_port = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.510372] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] backdoor_socket = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.510532] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] block_device_allocate_retries = 60 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.510689] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] block_device_allocate_retries_interval = 3 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.510853] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cert = self.pem {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.511027] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.511198] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute_monitors = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.511363] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] config_dir = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.511532] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] config_drive_format = iso9660 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.511666] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.511828] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] config_source = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.511994] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] console_host = devstack {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.512173] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] control_exchange = nova {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.512330] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cpu_allocation_ratio = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.512487] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] daemon = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.512653] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] debug = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.512808] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] default_access_ip_network_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.512971] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] default_availability_zone = nova {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.513141] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] default_ephemeral_format = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.513299] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] default_green_pool_size = 1000 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.513531] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.513694] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] default_schedule_zone = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.513850] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] disk_allocation_ratio = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.514040] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] enable_new_services = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.514231] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] enabled_apis = ['osapi_compute'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.514398] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] enabled_ssl_apis = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.514558] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] flat_injected = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.514715] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] force_config_drive = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.514865] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] force_raw_images = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.515048] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] graceful_shutdown_timeout = 5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.515212] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] heal_instance_info_cache_interval = 60 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.515422] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] host = cpu-1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.515597] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.515760] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] initial_disk_allocation_ratio = 1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.515921] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] initial_ram_allocation_ratio = 1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.516144] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.516329] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] instance_build_timeout = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.516504] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] instance_delete_interval = 300 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.516671] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] instance_format = [instance: %(uuid)s] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.516835] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] instance_name_template = instance-%08x {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.516997] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] instance_usage_audit = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.517185] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] instance_usage_audit_period = month {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.517351] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.517518] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] instances_path = /opt/stack/data/nova/instances {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.517683] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] internal_service_availability_zone = internal {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.517837] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] key = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.517996] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] live_migration_retry_count = 30 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.518178] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] log_color = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.518340] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] log_config_append = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.518503] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.518660] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] log_dir = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.518822] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] log_file = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.518942] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] log_options = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.519113] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] log_rotate_interval = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.519283] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] log_rotate_interval_type = days {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.519443] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] log_rotation_type = none {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.519571] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.519694] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.519857] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.520026] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.520157] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.520316] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] long_rpc_timeout = 1800 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.520474] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] max_concurrent_builds = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.520628] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] max_concurrent_live_migrations = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.520783] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] max_concurrent_snapshots = 5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.520937] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] max_local_block_devices = 3 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.521103] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] max_logfile_count = 30 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.521262] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] max_logfile_size_mb = 200 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.521417] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] maximum_instance_delete_attempts = 5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.521580] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] metadata_listen = 0.0.0.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.521744] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] metadata_listen_port = 8775 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.521909] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] metadata_workers = 2 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.522080] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] migrate_max_retries = -1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.522254] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] mkisofs_cmd = genisoimage {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.522454] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] my_block_storage_ip = 10.180.1.21 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.522586] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] my_ip = 10.180.1.21 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.522750] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] network_allocate_retries = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.522927] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.523103] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] osapi_compute_listen = 0.0.0.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.523269] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] osapi_compute_listen_port = 8774 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.523431] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] osapi_compute_unique_server_name_scope = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.523595] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] osapi_compute_workers = 2 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.523754] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] password_length = 12 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.523911] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] periodic_enable = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.524118] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] periodic_fuzzy_delay = 60 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.524303] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] pointer_model = usbtablet {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.524467] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] preallocate_images = none {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.524626] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] publish_errors = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.524757] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] pybasedir = /opt/stack/nova {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.524913] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ram_allocation_ratio = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.525132] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] rate_limit_burst = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.525317] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] rate_limit_except_level = CRITICAL {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.525479] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] rate_limit_interval = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.525639] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] reboot_timeout = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.525795] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] reclaim_instance_interval = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.525951] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] record = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.526133] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] reimage_timeout_per_gb = 60 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.526311] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] report_interval = 120 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.526472] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] rescue_timeout = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.526628] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] reserved_host_cpus = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.526785] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] reserved_host_disk_mb = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.526941] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] reserved_host_memory_mb = 512 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.527113] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] reserved_huge_pages = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.527274] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] resize_confirm_window = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.527429] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] resize_fs_using_block_device = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.527584] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] resume_guests_state_on_host_boot = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.527748] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.527906] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] rpc_response_timeout = 60 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.528073] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] run_external_periodic_tasks = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.528242] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] running_deleted_instance_action = reap {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.528399] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] running_deleted_instance_poll_interval = 1800 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.528554] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] running_deleted_instance_timeout = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.528710] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] scheduler_instance_sync_interval = 120 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.528874] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] service_down_time = 720 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.529048] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] servicegroup_driver = db {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.529210] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] shell_completion = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.529368] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] shelved_offload_time = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.529525] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] shelved_poll_interval = 3600 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.529688] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] shutdown_timeout = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.529843] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] source_is_ipv6 = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.529996] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ssl_only = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.530249] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.530415] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] sync_power_state_interval = 600 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.530577] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] sync_power_state_pool_size = 1000 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.530744] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] syslog_log_facility = LOG_USER {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.530897] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] tempdir = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.531065] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] timeout_nbd = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.531239] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] transport_url = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.531402] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] update_resources_interval = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.531556] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] use_cow_images = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.531712] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] use_eventlog = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.531868] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] use_journal = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.532033] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] use_json = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.532195] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] use_rootwrap_daemon = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.532352] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] use_stderr = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.532510] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] use_syslog = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.532666] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vcpu_pin_set = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.532831] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vif_plugging_is_fatal = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.532998] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vif_plugging_timeout = 300 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.533178] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] virt_mkfs = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.533340] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] volume_usage_poll_interval = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.533499] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] watch_log_file = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.533664] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] web = /usr/share/spice-html5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 478.533844] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_concurrency.disable_process_locking = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.534462] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.534703] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.534887] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.535089] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.535290] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.535464] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.535647] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.auth_strategy = keystone {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.535816] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.compute_link_prefix = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.535991] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.536184] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.dhcp_domain = novalocal {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.536355] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.enable_instance_password = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.536519] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.glance_link_prefix = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.536686] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.536856] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.537033] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.instance_list_per_project_cells = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.537203] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.list_records_by_skipping_down_cells = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.537367] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.local_metadata_per_cell = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.537534] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.max_limit = 1000 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.537702] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.metadata_cache_expiration = 15 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.537872] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.neutron_default_tenant_id = default {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.538053] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.response_validation = warn {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.538227] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.use_neutron_default_nets = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.538394] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.538556] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.538725] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.538896] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.539115] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.vendordata_dynamic_targets = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.539317] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.vendordata_jsonfile_path = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.539505] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.539701] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.backend = dogpile.cache.memcached {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.539870] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.backend_argument = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.540054] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.config_prefix = cache.oslo {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.540232] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.dead_timeout = 60.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.540399] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.debug_cache_backend = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.540562] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.enable_retry_client = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.540724] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.enable_socket_keepalive = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.540895] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.enabled = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.541071] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.enforce_fips_mode = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.541243] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.expiration_time = 600 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.541405] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.hashclient_retry_attempts = 2 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.541570] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.hashclient_retry_delay = 1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.541732] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.memcache_dead_retry = 300 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.541891] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.memcache_password = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.542063] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.542307] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.542505] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.memcache_pool_maxsize = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.542676] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.542841] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.memcache_sasl_enabled = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.543031] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.543210] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.memcache_socket_timeout = 1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.543373] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.memcache_username = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.543542] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.proxies = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.543706] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.redis_db = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.543865] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.redis_password = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.544071] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.redis_sentinel_service_name = mymaster {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.544264] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.544435] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.redis_server = localhost:6379 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.544600] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.redis_socket_timeout = 1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.544757] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.redis_username = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.544918] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.retry_attempts = 2 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.545141] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.retry_delay = 0.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.545323] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.socket_keepalive_count = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.545485] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.socket_keepalive_idle = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.545646] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.socket_keepalive_interval = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.545803] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.tls_allowed_ciphers = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.545959] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.tls_cafile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.546130] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.tls_certfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.546292] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.tls_enabled = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.546447] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cache.tls_keyfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.546615] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.auth_section = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.546788] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.auth_type = password {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.546948] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.cafile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.547135] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.catalog_info = volumev3::publicURL {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.547299] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.certfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.547460] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.collect_timing = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.547620] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.cross_az_attach = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.547780] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.debug = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.547939] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.endpoint_template = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.548116] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.http_retries = 3 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.548285] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.insecure = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.548443] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.keyfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.548612] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.os_region_name = RegionOne {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.548773] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.split_loggers = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.548933] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cinder.timeout = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.549119] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.549285] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute.cpu_dedicated_set = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.549445] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute.cpu_shared_set = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.549612] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute.image_type_exclude_list = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.549777] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.549941] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute.max_concurrent_disk_ops = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.550116] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute.max_disk_devices_to_attach = -1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.550284] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.550454] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.550619] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute.resource_provider_association_refresh = 300 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.550781] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.550944] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute.shutdown_retry_interval = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.551139] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.551323] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] conductor.workers = 2 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.551502] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] console.allowed_origins = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.551666] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] console.ssl_ciphers = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.551838] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] console.ssl_minimum_version = default {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.552015] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] consoleauth.enforce_session_timeout = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.552198] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] consoleauth.token_ttl = 600 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.552374] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.cafile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.552535] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.certfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.552696] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.collect_timing = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.552855] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.connect_retries = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.553028] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.connect_retry_delay = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.553193] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.endpoint_override = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.553356] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.insecure = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.553512] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.keyfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.553672] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.max_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.553830] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.min_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.554018] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.region_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.554196] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.retriable_status_codes = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.554358] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.service_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.554527] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.service_type = accelerator {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.554687] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.split_loggers = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.554845] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.status_code_retries = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.555028] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.status_code_retry_delay = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.555229] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.timeout = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.555420] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.555583] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] cyborg.version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.555761] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.backend = sqlalchemy {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.555930] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.connection = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.556108] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.connection_debug = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.556285] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.connection_parameters = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.556446] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.connection_recycle_time = 3600 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.556607] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.connection_trace = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.556769] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.db_inc_retry_interval = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.556926] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.db_max_retries = 20 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.557098] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.db_max_retry_interval = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.557263] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.db_retry_interval = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.557421] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.max_overflow = 50 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.557581] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.max_pool_size = 5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.557739] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.max_retries = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.557904] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.558073] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.mysql_wsrep_sync_wait = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.558234] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.pool_timeout = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.558394] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.retry_interval = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.558551] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.slave_connection = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.558711] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.sqlite_synchronous = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.558871] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] database.use_db_reconnect = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.559059] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.backend = sqlalchemy {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.559238] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.connection = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.559400] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.connection_debug = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.559570] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.connection_parameters = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.559729] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.connection_recycle_time = 3600 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.559889] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.connection_trace = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.560060] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.db_inc_retry_interval = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.560229] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.db_max_retries = 20 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.560388] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.db_max_retry_interval = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.560548] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.db_retry_interval = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.560711] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.max_overflow = 50 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.560870] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.max_pool_size = 5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.561039] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.max_retries = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.561214] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.561372] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.561529] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.pool_timeout = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.561690] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.retry_interval = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.561847] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.slave_connection = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.562017] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] api_database.sqlite_synchronous = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.562204] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] devices.enabled_mdev_types = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.562380] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.562548] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ephemeral_storage_encryption.default_format = luks {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.562708] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ephemeral_storage_encryption.enabled = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.562869] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.563049] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.api_servers = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.563221] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.cafile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.563382] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.certfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.563543] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.collect_timing = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.563700] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.connect_retries = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.563857] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.connect_retry_delay = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.564054] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.debug = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.564238] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.default_trusted_certificate_ids = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.564401] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.enable_certificate_validation = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.564563] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.enable_rbd_download = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.564718] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.endpoint_override = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.564881] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.insecure = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.565084] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.keyfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.565274] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.max_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.565435] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.min_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.565597] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.num_retries = 3 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.565767] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.rbd_ceph_conf = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.565932] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.rbd_connect_timeout = 5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.566114] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.rbd_pool = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.566283] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.rbd_user = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.566442] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.region_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.566602] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.retriable_status_codes = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.566757] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.service_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.566922] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.service_type = image {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.567095] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.split_loggers = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.567259] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.status_code_retries = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.567417] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.status_code_retry_delay = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.567573] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.timeout = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.567758] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.567923] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.verify_glance_signatures = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.568092] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] glance.version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.568262] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] guestfs.debug = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.568429] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] mks.enabled = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.568781] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.568970] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] image_cache.manager_interval = 2400 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.569156] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] image_cache.precache_concurrency = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.569324] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] image_cache.remove_unused_base_images = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.569491] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.569655] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.569827] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] image_cache.subdirectory_name = _base {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.570009] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.api_max_retries = 60 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.570185] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.api_retry_interval = 2 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.570343] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.auth_section = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.570504] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.auth_type = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.570663] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.cafile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.571330] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.certfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.571507] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.collect_timing = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.571674] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.conductor_group = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.571832] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.connect_retries = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.571993] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.connect_retry_delay = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.572166] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.endpoint_override = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.572328] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.insecure = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.572488] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.keyfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.572646] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.max_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.572805] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.min_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.572970] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.peer_list = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.573144] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.region_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.573304] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.retriable_status_codes = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.573466] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.serial_console_state_timeout = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.573624] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.service_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.573791] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.service_type = baremetal {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.573950] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.shard = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.574156] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.split_loggers = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.574323] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.status_code_retries = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.574482] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.status_code_retry_delay = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.574639] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.timeout = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.574818] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.574998] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ironic.version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.575229] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.575414] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] key_manager.fixed_key = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.575598] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.575762] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.barbican_api_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.575923] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.barbican_endpoint = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.576107] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.barbican_endpoint_type = public {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.576274] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.barbican_region_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.576436] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.cafile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.576595] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.certfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.576755] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.collect_timing = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.576914] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.insecure = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.577081] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.keyfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.577251] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.number_of_retries = 60 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.577413] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.retry_delay = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.577573] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.send_service_user_token = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.577737] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.split_loggers = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.577895] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.timeout = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.578067] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.verify_ssl = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.578230] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican.verify_ssl_path = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.578398] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican_service_user.auth_section = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.578562] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican_service_user.auth_type = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.578720] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican_service_user.cafile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.578873] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican_service_user.certfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.579044] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican_service_user.collect_timing = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.579212] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican_service_user.insecure = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.579367] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican_service_user.keyfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.579527] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican_service_user.split_loggers = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.579682] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] barbican_service_user.timeout = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.579847] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vault.approle_role_id = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.580023] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vault.approle_secret_id = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.580195] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vault.kv_mountpoint = secret {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.580355] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vault.kv_path = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.580519] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vault.kv_version = 2 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.580678] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vault.namespace = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.580834] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vault.root_token_id = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.580992] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vault.ssl_ca_crt_file = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.581173] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vault.timeout = 60.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.581333] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vault.use_ssl = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.581500] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.581667] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.auth_section = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.581828] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.auth_type = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.581996] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.cafile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.582160] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.certfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.582322] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.collect_timing = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.582478] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.connect_retries = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.582633] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.connect_retry_delay = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.582788] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.endpoint_override = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.582946] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.insecure = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.583121] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.keyfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.583281] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.max_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.583434] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.min_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.583589] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.region_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.583745] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.retriable_status_codes = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.583900] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.service_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.584101] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.service_type = identity {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.584275] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.split_loggers = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.584436] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.status_code_retries = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.584593] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.status_code_retry_delay = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.584749] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.timeout = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.584927] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.585114] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] keystone.version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.585329] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.connection_uri = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.585493] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.cpu_mode = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.585660] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.cpu_model_extra_flags = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.585829] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.cpu_models = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.585999] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.cpu_power_governor_high = performance {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.586186] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.cpu_power_governor_low = powersave {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.586351] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.cpu_power_management = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.586519] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.586684] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.device_detach_attempts = 8 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.586847] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.device_detach_timeout = 20 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.587016] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.disk_cachemodes = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.587180] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.disk_prefix = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.587347] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.enabled_perf_events = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.587512] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.file_backed_memory = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.587676] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.gid_maps = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.587835] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.hw_disk_discard = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.587991] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.hw_machine_type = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.588175] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.images_rbd_ceph_conf = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.588339] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.588499] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.588667] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.images_rbd_glance_store_name = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.588836] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.images_rbd_pool = rbd {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.589010] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.images_type = default {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.589178] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.images_volume_group = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.589339] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.inject_key = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.589499] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.inject_partition = -2 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.589655] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.inject_password = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.589814] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.iscsi_iface = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.589974] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.iser_use_multipath = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.590154] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.live_migration_bandwidth = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.590319] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.590480] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.live_migration_downtime = 500 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.590640] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.590800] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.590958] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.live_migration_inbound_addr = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.591130] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.591294] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.live_migration_permit_post_copy = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.591447] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.live_migration_scheme = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.591613] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.live_migration_timeout_action = abort {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.591780] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.live_migration_tunnelled = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.591941] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.live_migration_uri = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.592120] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.live_migration_with_native_tls = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.592287] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.max_queues = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.592452] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.592683] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.592847] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.nfs_mount_options = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.593154] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.593333] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.593501] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.num_iser_scan_tries = 5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.593664] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.num_memory_encrypted_guests = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.593829] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.594027] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.num_pcie_ports = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.594210] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.num_volume_scan_tries = 5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.594378] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.pmem_namespaces = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.594537] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.quobyte_client_cfg = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.594819] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.595012] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.rbd_connect_timeout = 5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.595225] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.595397] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.595556] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.rbd_secret_uuid = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.595713] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.rbd_user = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.595875] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.596054] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.remote_filesystem_transport = ssh {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.596222] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.rescue_image_id = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.596379] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.rescue_kernel_id = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.596532] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.rescue_ramdisk_id = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.596697] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.596852] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.rx_queue_size = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.597027] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.smbfs_mount_options = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.597306] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.597478] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.snapshot_compression = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.597637] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.snapshot_image_format = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.597854] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.598030] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.sparse_logical_volumes = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.598200] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.swtpm_enabled = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.598369] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.swtpm_group = tss {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.598533] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.swtpm_user = tss {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.598699] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.sysinfo_serial = unique {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.598856] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.tb_cache_size = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.599021] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.tx_queue_size = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.599192] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.uid_maps = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.599355] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.use_virtio_for_bridges = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.599523] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.virt_type = kvm {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.599688] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.volume_clear = zero {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.599849] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.volume_clear_size = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.600021] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.volume_use_multipath = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.600184] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.vzstorage_cache_path = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.600351] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.600515] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.vzstorage_mount_group = qemu {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.600676] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.vzstorage_mount_opts = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.600840] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.601120] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.601298] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.vzstorage_mount_user = stack {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.601463] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.601635] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.auth_section = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.601806] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.auth_type = password {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.601967] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.cafile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.602141] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.certfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.602303] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.collect_timing = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.602459] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.connect_retries = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.602616] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.connect_retry_delay = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.602784] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.default_floating_pool = public {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.602943] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.endpoint_override = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.603116] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.extension_sync_interval = 600 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.603307] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.http_retries = 3 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.603488] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.insecure = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.603649] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.keyfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.603807] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.max_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.603982] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.604175] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.min_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.604349] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.ovs_bridge = br-int {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.604513] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.physnets = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.604680] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.region_name = RegionOne {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.604839] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.retriable_status_codes = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.605019] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.service_metadata_proxy = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.605201] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.service_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.605373] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.service_type = network {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.605600] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.split_loggers = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.605780] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.status_code_retries = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.605941] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.status_code_retry_delay = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.606113] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.timeout = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.606305] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.606468] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] neutron.version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.606638] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] notifications.bdms_in_notifications = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.606812] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] notifications.default_level = INFO {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.606997] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] notifications.notification_format = unversioned {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.607178] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] notifications.notify_on_state_change = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.607355] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.607529] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] pci.alias = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.607696] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] pci.device_spec = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.607857] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] pci.report_in_placement = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.608035] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.auth_section = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.608215] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.auth_type = password {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.608383] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.608543] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.cafile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.608700] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.certfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.608858] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.collect_timing = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.609021] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.connect_retries = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.609188] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.connect_retry_delay = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.609344] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.default_domain_id = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.609498] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.default_domain_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.609652] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.domain_id = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.609805] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.domain_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.609959] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.endpoint_override = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.610130] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.insecure = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.610290] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.keyfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.610444] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.max_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.610597] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.min_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.610762] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.password = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.610920] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.project_domain_id = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.611096] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.project_domain_name = Default {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.611267] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.project_id = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.611436] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.project_name = service {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.611603] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.region_name = RegionOne {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.611766] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.retriable_status_codes = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.611922] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.service_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.612105] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.service_type = placement {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.612268] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.split_loggers = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.612425] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.status_code_retries = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.612585] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.status_code_retry_delay = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.612743] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.system_scope = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.612899] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.timeout = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.613064] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.trust_id = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.613227] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.user_domain_id = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.613393] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.user_domain_name = Default {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.613550] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.user_id = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.613720] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.username = nova {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.613897] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.614088] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] placement.version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.614278] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] quota.cores = 20 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.614443] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] quota.count_usage_from_placement = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.614611] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.614785] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] quota.injected_file_content_bytes = 10240 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.614952] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] quota.injected_file_path_length = 255 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.615183] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] quota.injected_files = 5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.615364] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] quota.instances = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.615529] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] quota.key_pairs = 100 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.615693] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] quota.metadata_items = 128 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.615857] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] quota.ram = 51200 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.616027] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] quota.recheck_quota = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.616198] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] quota.server_group_members = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.616363] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] quota.server_groups = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.616537] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.616702] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.616862] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] scheduler.image_metadata_prefilter = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.617032] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.617201] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] scheduler.max_attempts = 3 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.617363] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] scheduler.max_placement_results = 1000 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.617522] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.617681] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] scheduler.query_placement_for_image_type_support = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.617840] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.618021] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] scheduler.workers = 2 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.618196] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.618362] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.618537] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.618703] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.618865] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.619036] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.619206] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.619388] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.619553] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.host_subset_size = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.619713] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.619869] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.620039] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.620211] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.isolated_hosts = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.620379] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.isolated_images = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.620538] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.620697] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.620857] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.621028] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.pci_in_placement = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.621197] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.621356] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.621514] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.621671] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.621829] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.621990] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.622164] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.track_instance_changes = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.622338] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.622505] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] metrics.required = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.622667] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] metrics.weight_multiplier = 1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.622830] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.622994] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] metrics.weight_setting = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.623325] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.623499] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] serial_console.enabled = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.623673] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] serial_console.port_range = 10000:20000 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.623840] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.624034] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.624219] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] serial_console.serialproxy_port = 6083 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.624388] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] service_user.auth_section = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.624559] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] service_user.auth_type = password {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.624720] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] service_user.cafile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.624876] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] service_user.certfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.625050] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] service_user.collect_timing = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.625240] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] service_user.insecure = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.625401] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] service_user.keyfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.625573] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] service_user.send_service_user_token = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.625737] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] service_user.split_loggers = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.625896] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] service_user.timeout = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.626077] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] spice.agent_enabled = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.626246] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] spice.enabled = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.626571] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.626760] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.626928] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] spice.html5proxy_port = 6082 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.627101] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] spice.image_compression = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.627265] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] spice.jpeg_compression = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.627424] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] spice.playback_compression = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.627615] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] spice.require_secure = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.627833] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] spice.server_listen = 127.0.0.1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.628019] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.628184] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] spice.streaming_mode = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.628342] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] spice.zlib_compression = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.628510] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] upgrade_levels.baseapi = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.628678] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] upgrade_levels.compute = auto {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.628837] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] upgrade_levels.conductor = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.628993] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] upgrade_levels.scheduler = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.629205] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vendordata_dynamic_auth.auth_section = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.629392] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vendordata_dynamic_auth.auth_type = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.629555] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vendordata_dynamic_auth.cafile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.629714] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vendordata_dynamic_auth.certfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.629876] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.630046] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vendordata_dynamic_auth.insecure = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.630210] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vendordata_dynamic_auth.keyfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.630371] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.630527] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vendordata_dynamic_auth.timeout = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.630698] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.api_retry_count = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.630855] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.ca_file = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.631033] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.cache_prefix = devstack-image-cache {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.631205] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.cluster_name = testcl1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.631368] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.connection_pool_size = 10 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.631525] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.console_delay_seconds = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.631690] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.datastore_regex = ^datastore.* {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.631893] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.632080] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.host_password = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.632250] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.host_port = 443 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.632420] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.host_username = administrator@vsphere.local {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.632590] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.insecure = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.632750] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.integration_bridge = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.632912] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.maximum_objects = 100 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.633082] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.pbm_default_policy = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.633247] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.pbm_enabled = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.633404] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.pbm_wsdl_location = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.633569] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.633726] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.serial_port_proxy_uri = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.633881] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.serial_port_service_uri = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.634074] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.task_poll_interval = 0.5 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.634256] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.use_linked_clone = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.634424] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.vnc_keymap = en-us {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.634587] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.vnc_port = 5900 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.634747] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vmware.vnc_port_total = 10000 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.634931] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vnc.auth_schemes = ['none'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.635133] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vnc.enabled = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.635442] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.635626] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.635796] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vnc.novncproxy_port = 6080 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.635972] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vnc.server_listen = 127.0.0.1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.636162] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.636322] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vnc.vencrypt_ca_certs = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.636478] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vnc.vencrypt_client_cert = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.636634] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vnc.vencrypt_client_key = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.636814] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.636978] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.disable_deep_image_inspection = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.637159] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.637319] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.637479] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.637639] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.disable_rootwrap = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.637798] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.enable_numa_live_migration = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.637955] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.638127] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.638292] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.638451] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.libvirt_disable_apic = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.638606] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.638764] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.638922] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.639091] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.639255] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.639413] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.639570] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.639723] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.639878] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.640049] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.640237] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.640406] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] wsgi.client_socket_timeout = 900 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.640571] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] wsgi.default_pool_size = 1000 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.640734] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] wsgi.keep_alive = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.640899] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] wsgi.max_header_line = 16384 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.641068] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] wsgi.secure_proxy_ssl_header = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.641233] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] wsgi.ssl_ca_file = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.641392] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] wsgi.ssl_cert_file = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.641551] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] wsgi.ssl_key_file = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.641712] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] wsgi.tcp_keepidle = 600 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.641893] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.642074] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] zvm.ca_file = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.642238] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] zvm.cloud_connector_url = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.642519] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.642692] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] zvm.reachable_timeout = 300 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.642873] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_policy.enforce_new_defaults = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.643283] env[63372]: WARNING oslo_config.cfg [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 478.643472] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_policy.enforce_scope = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.643647] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_policy.policy_default_rule = default {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.643826] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.644028] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_policy.policy_file = policy.yaml {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.644229] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.644396] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.644557] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.644715] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.644877] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.645062] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.645272] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.645451] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] profiler.connection_string = messaging:// {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.645621] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] profiler.enabled = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.645788] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] profiler.es_doc_type = notification {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.645950] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] profiler.es_scroll_size = 10000 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.646132] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] profiler.es_scroll_time = 2m {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.646306] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] profiler.filter_error_trace = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.646494] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] profiler.hmac_keys = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.646663] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] profiler.sentinel_service_name = mymaster {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.646830] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] profiler.socket_timeout = 0.1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.646991] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] profiler.trace_requests = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.647166] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] profiler.trace_sqlalchemy = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.647350] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] profiler_jaeger.process_tags = {} {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.647509] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] profiler_jaeger.service_name_prefix = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.647667] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] profiler_otlp.service_name_prefix = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.647831] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] remote_debug.host = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.647989] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] remote_debug.port = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.648184] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.648347] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.648507] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.648670] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.648825] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.648982] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.649157] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.649318] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.649476] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.649643] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.649800] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.649967] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.650148] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.650316] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.650479] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.650642] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.650800] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.650970] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.651145] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.651305] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.651465] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.651625] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.651783] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.651944] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.652115] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.652276] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.652435] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.652593] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.652755] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.652918] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.ssl = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.653100] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.653272] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.653433] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.653600] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.653767] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.ssl_version = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.653926] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.654157] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.654333] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_notifications.retry = -1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.654517] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.654690] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_messaging_notifications.transport_url = **** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.654860] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.auth_section = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.655032] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.auth_type = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.655224] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.cafile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.655386] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.certfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.655546] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.collect_timing = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.655703] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.connect_retries = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.655858] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.connect_retry_delay = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.656023] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.endpoint_id = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.656185] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.endpoint_override = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.656366] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.insecure = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.656532] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.keyfile = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.656691] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.max_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.656846] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.min_version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.656999] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.region_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.657179] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.retriable_status_codes = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.657336] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.service_name = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.657492] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.service_type = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.657651] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.split_loggers = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.657808] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.status_code_retries = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.657966] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.status_code_retry_delay = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.658137] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.timeout = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.658296] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.valid_interfaces = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.658453] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_limit.version = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.658614] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_reports.file_event_handler = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.658776] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.658989] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] oslo_reports.log_dir = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.659263] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.659489] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.659698] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.659875] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.660054] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.660222] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.660394] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.660555] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vif_plug_ovs_privileged.group = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.660714] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.660881] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.661056] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.661221] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] vif_plug_ovs_privileged.user = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.661394] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_linux_bridge.flat_interface = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.661574] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.661746] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.661917] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.662098] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.662271] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.662436] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.662597] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.662774] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.662945] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_ovs.isolate_vif = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.663123] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.663293] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.663461] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.663629] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_ovs.ovsdb_interface = native {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.663791] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_vif_ovs.per_port_bridge = False {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.663960] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_brick.lock_path = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.664169] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.664338] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.664508] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] privsep_osbrick.capabilities = [21] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.664667] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] privsep_osbrick.group = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.664824] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] privsep_osbrick.helper_command = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.664991] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.665189] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.665357] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] privsep_osbrick.user = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.665534] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.665696] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] nova_sys_admin.group = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.665854] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] nova_sys_admin.helper_command = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.666030] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.666201] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.666386] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] nova_sys_admin.user = None {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.666528] env[63372]: DEBUG oslo_service.service [None req-4e7ba833-17d5-48d7-bf8d-7099e57945b6 None None] ******************************************************************************** {{(pid=63372) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 478.667037] env[63372]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 479.170452] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Getting list of instances from cluster (obj){ [ 479.170452] env[63372]: value = "domain-c8" [ 479.170452] env[63372]: _type = "ClusterComputeResource" [ 479.170452] env[63372]: } {{(pid=63372) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 479.171621] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6384dc75-36fc-4257-90d2-14c6c0b3dcd9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 479.180905] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Got total of 0 instances {{(pid=63372) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 479.181445] env[63372]: WARNING nova.virt.vmwareapi.driver [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 479.181920] env[63372]: INFO nova.virt.node [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Generated node identity c1b110ca-a185-44c2-ba1c-f05b3b420add [ 479.182166] env[63372]: INFO nova.virt.node [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Wrote node identity c1b110ca-a185-44c2-ba1c-f05b3b420add to /opt/stack/data/n-cpu-1/compute_id [ 479.685024] env[63372]: WARNING nova.compute.manager [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Compute nodes ['c1b110ca-a185-44c2-ba1c-f05b3b420add'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 480.691280] env[63372]: INFO nova.compute.manager [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 481.696913] env[63372]: WARNING nova.compute.manager [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 481.697252] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 481.697430] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 481.697582] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 481.697737] env[63372]: DEBUG nova.compute.resource_tracker [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63372) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 481.698730] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90ab03d-a45f-4741-b6cd-21f848b444d1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.707346] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a669a01-bae5-48d2-b862-1cb1d861e803 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.720724] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d12182-3b35-402f-85db-107b8f8afe53 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.727101] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f61f870-aaf7-4b51-a943-d75819a43423 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.756079] env[63372]: DEBUG nova.compute.resource_tracker [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181573MB free_disk=186GB free_vcpus=48 pci_devices=None {{(pid=63372) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 481.756309] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 481.756551] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 482.258621] env[63372]: WARNING nova.compute.resource_tracker [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] No compute node record for cpu-1:c1b110ca-a185-44c2-ba1c-f05b3b420add: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host c1b110ca-a185-44c2-ba1c-f05b3b420add could not be found. [ 482.764905] env[63372]: INFO nova.compute.resource_tracker [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: c1b110ca-a185-44c2-ba1c-f05b3b420add [ 484.275720] env[63372]: DEBUG nova.compute.resource_tracker [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 484.276060] env[63372]: DEBUG nova.compute.resource_tracker [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 484.433119] env[63372]: INFO nova.scheduler.client.report [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] [req-6ffbffc7-6763-4ce6-9556-346699f738e0] Created resource provider record via placement API for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 484.449700] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d089a0f-1725-4414-83bf-3e86fa12205c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 484.457347] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0af366b-65bd-489e-946f-1ebc0633ff39 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 484.487324] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10e8611-fb2d-4e67-9594-fcbbd827baa1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 484.494250] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf77dbea-3ffb-4947-9729-014f8b058a60 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 484.506829] env[63372]: DEBUG nova.compute.provider_tree [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 485.048184] env[63372]: DEBUG nova.scheduler.client.report [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 485.048423] env[63372]: DEBUG nova.compute.provider_tree [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 0 to 1 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 485.048564] env[63372]: DEBUG nova.compute.provider_tree [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 485.096277] env[63372]: DEBUG nova.compute.provider_tree [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 1 to 2 during operation: update_traits {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 485.600991] env[63372]: DEBUG nova.compute.resource_tracker [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 485.601418] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.845s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 485.601418] env[63372]: DEBUG nova.service [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Creating RPC server for service compute {{(pid=63372) start /opt/stack/nova/nova/service.py:186}} [ 485.615806] env[63372]: DEBUG nova.service [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] Join ServiceGroup membership for this service compute {{(pid=63372) start /opt/stack/nova/nova/service.py:203}} [ 485.616013] env[63372]: DEBUG nova.servicegroup.drivers.db [None req-1b016ff4-db1f-436a-a186-80a02bc16e7e None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=63372) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 514.617558] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._sync_power_states {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 515.120372] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Getting list of instances from cluster (obj){ [ 515.120372] env[63372]: value = "domain-c8" [ 515.120372] env[63372]: _type = "ClusterComputeResource" [ 515.120372] env[63372]: } {{(pid=63372) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 515.121690] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eed6456-e386-4ee9-a189-8632948a073e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.131152] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Got total of 0 instances {{(pid=63372) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 515.131412] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 515.131708] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Getting list of instances from cluster (obj){ [ 515.131708] env[63372]: value = "domain-c8" [ 515.131708] env[63372]: _type = "ClusterComputeResource" [ 515.131708] env[63372]: } {{(pid=63372) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 515.132616] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee004fc-dc4a-4563-bc99-7e6bfa37ce05 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.140140] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Got total of 0 instances {{(pid=63372) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 519.529026] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Acquiring lock "5cf85718-bb73-4534-9dd9-ddb4b6777f6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.529952] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Lock "5cf85718-bb73-4534-9dd9-ddb4b6777f6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.032907] env[63372]: DEBUG nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 520.345146] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Acquiring lock "f33eea3e-887c-4164-98e9-e4d502d14af2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.345146] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Lock "f33eea3e-887c-4164-98e9-e4d502d14af2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.585316] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.585579] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.587341] env[63372]: INFO nova.compute.claims [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 520.853430] env[63372]: DEBUG nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 521.382128] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.654196] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5713382d-75e7-4853-9146-841568764d40 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.664294] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d5c816-27cc-470c-8090-d4ad1e2fe718 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.697165] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3320deac-ba0f-4652-81e7-5784151a216d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.705465] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d75fb2b-0556-40f7-9363-40b6ce5362e7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.722679] env[63372]: DEBUG nova.compute.provider_tree [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 522.113738] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Acquiring lock "393f3218-5e22-4e88-8968-9157aa0631da" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.114197] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Lock "393f3218-5e22-4e88-8968-9157aa0631da" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.226111] env[63372]: DEBUG nova.scheduler.client.report [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 522.619916] env[63372]: DEBUG nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 522.732408] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.147s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.733086] env[63372]: DEBUG nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 522.739808] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.357s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.740012] env[63372]: INFO nova.compute.claims [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 523.150611] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.239850] env[63372]: DEBUG nova.compute.utils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 523.243341] env[63372]: DEBUG nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 523.243600] env[63372]: DEBUG nova.network.neutron [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 523.558381] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Acquiring lock "32dabedb-8155-4377-871b-408935db97b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.560248] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Lock "32dabedb-8155-4377-871b-408935db97b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.661044] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "21c4415c-a016-47eb-9ec4-280c6c18b3fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.661302] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "21c4415c-a016-47eb-9ec4-280c6c18b3fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.737322] env[63372]: DEBUG nova.policy [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1625f9903b84460a8ee0318dda7adaad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b0dabe1be5324f31b3d7ee3d40352950', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 523.749218] env[63372]: DEBUG nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 523.852223] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523a2832-d0dd-459d-a01b-83c7612fc72d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.861531] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc20dcd-08bb-473a-ad98-ee121458916e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.896457] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7079d0c0-81c9-4ff8-b967-e4770857b547 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.903994] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58baaf14-2d42-46aa-af75-fb5fa379465c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.917477] env[63372]: DEBUG nova.compute.provider_tree [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 524.062838] env[63372]: DEBUG nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 524.166022] env[63372]: DEBUG nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 524.423098] env[63372]: DEBUG nova.scheduler.client.report [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 524.587337] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.692596] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.758666] env[63372]: DEBUG nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 524.761775] env[63372]: DEBUG nova.network.neutron [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Successfully created port: 8c7ae2fc-56c6-43bd-8f50-5a1eebe4fa92 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 524.783715] env[63372]: DEBUG nova.virt.hardware [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 524.784423] env[63372]: DEBUG nova.virt.hardware [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 524.784423] env[63372]: DEBUG nova.virt.hardware [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 524.784423] env[63372]: DEBUG nova.virt.hardware [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 524.784588] env[63372]: DEBUG nova.virt.hardware [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 524.784720] env[63372]: DEBUG nova.virt.hardware [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 524.785031] env[63372]: DEBUG nova.virt.hardware [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 524.785179] env[63372]: DEBUG nova.virt.hardware [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 524.785558] env[63372]: DEBUG nova.virt.hardware [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 524.785721] env[63372]: DEBUG nova.virt.hardware [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 524.785913] env[63372]: DEBUG nova.virt.hardware [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 524.787283] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae918ff-9cb6-4e63-93b5-34616efccca3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.795778] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef50af4-6b2d-4554-beb8-fdb0f0e93560 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.811410] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed453ab-2e26-4410-8feb-e503ca6f450f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.927048] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.188s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.927048] env[63372]: DEBUG nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 524.929339] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.781s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.931075] env[63372]: INFO nova.compute.claims [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 525.437203] env[63372]: DEBUG nova.compute.utils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 525.441930] env[63372]: DEBUG nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 525.442352] env[63372]: DEBUG nova.network.neutron [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 525.612979] env[63372]: DEBUG nova.policy [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '942942a0be914cd1859affad1332814e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7f3c1e284ca9434fb7d57ab3fb54fa39', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 525.946634] env[63372]: DEBUG nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 526.058015] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20271ce-b7bb-49ba-bd37-f90a43644f1f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.066225] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080d69a3-9022-4f01-9b10-d0d59709d739 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.098933] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29c639ac-d376-400e-a603-73fcda648b46 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.107980] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95d40a4-32ca-4bcb-89da-ed86528c13e5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.123137] env[63372]: DEBUG nova.compute.provider_tree [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 526.631335] env[63372]: DEBUG nova.scheduler.client.report [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 526.959235] env[63372]: DEBUG nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 526.997719] env[63372]: DEBUG nova.virt.hardware [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 527.000022] env[63372]: DEBUG nova.virt.hardware [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 527.000022] env[63372]: DEBUG nova.virt.hardware [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 527.000022] env[63372]: DEBUG nova.virt.hardware [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 527.000022] env[63372]: DEBUG nova.virt.hardware [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 527.000022] env[63372]: DEBUG nova.virt.hardware [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 527.000218] env[63372]: DEBUG nova.virt.hardware [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 527.000218] env[63372]: DEBUG nova.virt.hardware [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 527.000218] env[63372]: DEBUG nova.virt.hardware [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 527.000218] env[63372]: DEBUG nova.virt.hardware [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 527.000218] env[63372]: DEBUG nova.virt.hardware [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 527.002827] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6020ff7-a272-424a-8a5d-737962cb50b1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.017823] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67d4b45-7752-4f41-a016-1108588b88f1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.138974] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.209s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.139265] env[63372]: DEBUG nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 527.143724] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.556s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.146651] env[63372]: INFO nova.compute.claims [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 527.461201] env[63372]: DEBUG nova.network.neutron [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Successfully created port: 6590d44b-a55c-49b4-8105-3915d4a6d5a2 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 527.655156] env[63372]: DEBUG nova.compute.utils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 527.655793] env[63372]: DEBUG nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 527.656161] env[63372]: DEBUG nova.network.neutron [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 527.900223] env[63372]: DEBUG nova.policy [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f393cb29aaf347d6aa72a965d0a2a818', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bde7bd2a98074a69a37c208b2e8a7b60', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 528.160930] env[63372]: DEBUG nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 528.275061] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9781982e-e4b9-4537-863e-9efd8aae65b2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.283242] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36887437-1e5e-4829-a224-096912edb17e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.315014] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76988ff6-e756-497c-9aec-fa56e9739300 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.322409] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d737fbf6-1e7e-4a7f-a8c7-de299d0833e2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.336367] env[63372]: DEBUG nova.compute.provider_tree [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.841625] env[63372]: DEBUG nova.scheduler.client.report [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 529.172296] env[63372]: DEBUG nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 529.207586] env[63372]: DEBUG nova.virt.hardware [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 529.207764] env[63372]: DEBUG nova.virt.hardware [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 529.207852] env[63372]: DEBUG nova.virt.hardware [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 529.209121] env[63372]: DEBUG nova.virt.hardware [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 529.209636] env[63372]: DEBUG nova.virt.hardware [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 529.210104] env[63372]: DEBUG nova.virt.hardware [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 529.210434] env[63372]: DEBUG nova.virt.hardware [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 529.210602] env[63372]: DEBUG nova.virt.hardware [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 529.210768] env[63372]: DEBUG nova.virt.hardware [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 529.210938] env[63372]: DEBUG nova.virt.hardware [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 529.211146] env[63372]: DEBUG nova.virt.hardware [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 529.213316] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7597f0a2-dd37-4e8e-8370-32c17934c447 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.229321] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91315423-96be-459d-a2be-9a2b92ec015e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.331657] env[63372]: ERROR nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8c7ae2fc-56c6-43bd-8f50-5a1eebe4fa92, please check neutron logs for more information. [ 529.331657] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 529.331657] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.331657] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 529.331657] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.331657] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 529.331657] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.331657] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 529.331657] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.331657] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 529.331657] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.331657] env[63372]: ERROR nova.compute.manager raise self.value [ 529.331657] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.331657] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 529.331657] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.331657] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 529.332217] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.332217] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 529.332217] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8c7ae2fc-56c6-43bd-8f50-5a1eebe4fa92, please check neutron logs for more information. [ 529.332217] env[63372]: ERROR nova.compute.manager [ 529.332217] env[63372]: Traceback (most recent call last): [ 529.332217] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 529.332217] env[63372]: listener.cb(fileno) [ 529.332217] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.332217] env[63372]: result = function(*args, **kwargs) [ 529.332217] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.332217] env[63372]: return func(*args, **kwargs) [ 529.332217] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.332217] env[63372]: raise e [ 529.332217] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.332217] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 529.332217] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.332217] env[63372]: created_port_ids = self._update_ports_for_instance( [ 529.332217] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.332217] env[63372]: with excutils.save_and_reraise_exception(): [ 529.332217] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.332217] env[63372]: self.force_reraise() [ 529.332217] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.332217] env[63372]: raise self.value [ 529.332217] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.332217] env[63372]: updated_port = self._update_port( [ 529.332217] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.332217] env[63372]: _ensure_no_port_binding_failure(port) [ 529.332217] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.332217] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 529.333051] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 8c7ae2fc-56c6-43bd-8f50-5a1eebe4fa92, please check neutron logs for more information. [ 529.333051] env[63372]: Removing descriptor: 14 [ 529.333051] env[63372]: ERROR nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8c7ae2fc-56c6-43bd-8f50-5a1eebe4fa92, please check neutron logs for more information. [ 529.333051] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Traceback (most recent call last): [ 529.333051] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 529.333051] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] yield resources [ 529.333051] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 529.333051] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] self.driver.spawn(context, instance, image_meta, [ 529.333051] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 529.333051] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 529.333051] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 529.333051] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] vm_ref = self.build_virtual_machine(instance, [ 529.333545] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 529.333545] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] vif_infos = vmwarevif.get_vif_info(self._session, [ 529.333545] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 529.333545] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] for vif in network_info: [ 529.333545] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 529.333545] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] return self._sync_wrapper(fn, *args, **kwargs) [ 529.333545] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 529.333545] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] self.wait() [ 529.333545] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 529.333545] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] self[:] = self._gt.wait() [ 529.333545] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 529.333545] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] return self._exit_event.wait() [ 529.333545] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 529.334992] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] result = hub.switch() [ 529.334992] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 529.334992] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] return self.greenlet.switch() [ 529.334992] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.334992] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] result = function(*args, **kwargs) [ 529.334992] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.334992] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] return func(*args, **kwargs) [ 529.334992] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.334992] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] raise e [ 529.334992] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.334992] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] nwinfo = self.network_api.allocate_for_instance( [ 529.334992] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.334992] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] created_port_ids = self._update_ports_for_instance( [ 529.335367] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.335367] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] with excutils.save_and_reraise_exception(): [ 529.335367] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.335367] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] self.force_reraise() [ 529.335367] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.335367] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] raise self.value [ 529.335367] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.335367] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] updated_port = self._update_port( [ 529.335367] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.335367] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] _ensure_no_port_binding_failure(port) [ 529.335367] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.335367] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] raise exception.PortBindingFailed(port_id=port['id']) [ 529.335657] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] nova.exception.PortBindingFailed: Binding failed for port 8c7ae2fc-56c6-43bd-8f50-5a1eebe4fa92, please check neutron logs for more information. [ 529.335657] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] [ 529.335657] env[63372]: INFO nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Terminating instance [ 529.338287] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Acquiring lock "refresh_cache-5cf85718-bb73-4534-9dd9-ddb4b6777f6e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.338287] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Acquired lock "refresh_cache-5cf85718-bb73-4534-9dd9-ddb4b6777f6e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.338494] env[63372]: DEBUG nova.network.neutron [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 529.347820] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.204s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.348585] env[63372]: DEBUG nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 529.352194] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.660s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.353957] env[63372]: INFO nova.compute.claims [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.512493] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Acquiring lock "91351ef7-a282-4f57-b96b-ba0af6d0e074" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.512493] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Lock "91351ef7-a282-4f57-b96b-ba0af6d0e074" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.517509] env[63372]: DEBUG nova.network.neutron [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Successfully created port: 5f466550-60f4-48bb-871d-060148bd4558 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 529.830086] env[63372]: DEBUG nova.compute.manager [req-d31e204b-df4f-46f8-9a91-12148e57afca req-3dc55255-e455-437a-83d2-c2f47e191444 service nova] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Received event network-changed-8c7ae2fc-56c6-43bd-8f50-5a1eebe4fa92 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.830086] env[63372]: DEBUG nova.compute.manager [req-d31e204b-df4f-46f8-9a91-12148e57afca req-3dc55255-e455-437a-83d2-c2f47e191444 service nova] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Refreshing instance network info cache due to event network-changed-8c7ae2fc-56c6-43bd-8f50-5a1eebe4fa92. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 529.830281] env[63372]: DEBUG oslo_concurrency.lockutils [req-d31e204b-df4f-46f8-9a91-12148e57afca req-3dc55255-e455-437a-83d2-c2f47e191444 service nova] Acquiring lock "refresh_cache-5cf85718-bb73-4534-9dd9-ddb4b6777f6e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.862151] env[63372]: DEBUG nova.compute.utils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 529.864798] env[63372]: DEBUG nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 529.864798] env[63372]: DEBUG nova.network.neutron [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 529.908030] env[63372]: DEBUG nova.network.neutron [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 530.016510] env[63372]: DEBUG nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 530.024604] env[63372]: DEBUG nova.policy [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '485b88bf17f04f789145ef10e3a594eb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e21f8924c2446f0a32154d5e454aad8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 530.324090] env[63372]: DEBUG nova.network.neutron [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.369380] env[63372]: DEBUG nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 530.497110] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d370d36d-31fa-4b04-a2b0-51f34d23d7a4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.507637] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98605561-6819-4612-86e1-6f2aa50f9a96 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.551656] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282d9539-b9c1-478b-b8b4-531598cf9b3d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.560124] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3128c91-2387-4c7d-b078-d846f2c995e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.569320] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.578826] env[63372]: DEBUG nova.compute.provider_tree [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 530.827399] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Releasing lock "refresh_cache-5cf85718-bb73-4534-9dd9-ddb4b6777f6e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.827814] env[63372]: DEBUG nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 530.828027] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 530.828329] env[63372]: DEBUG oslo_concurrency.lockutils [req-d31e204b-df4f-46f8-9a91-12148e57afca req-3dc55255-e455-437a-83d2-c2f47e191444 service nova] Acquired lock "refresh_cache-5cf85718-bb73-4534-9dd9-ddb4b6777f6e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.828556] env[63372]: DEBUG nova.network.neutron [req-d31e204b-df4f-46f8-9a91-12148e57afca req-3dc55255-e455-437a-83d2-c2f47e191444 service nova] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Refreshing network info cache for port 8c7ae2fc-56c6-43bd-8f50-5a1eebe4fa92 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 530.829614] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bfafaa43-f5a8-4da1-8ea8-fcc98b8f41e9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.840231] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff3b60a-54bc-405b-86c4-a757783c21f2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.866111] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5cf85718-bb73-4534-9dd9-ddb4b6777f6e could not be found. [ 530.866111] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 530.866111] env[63372]: INFO nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 530.866111] env[63372]: DEBUG oslo.service.loopingcall [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 530.866111] env[63372]: DEBUG nova.compute.manager [-] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 530.866111] env[63372]: DEBUG nova.network.neutron [-] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 530.919078] env[63372]: DEBUG nova.network.neutron [-] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.084514] env[63372]: DEBUG nova.scheduler.client.report [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.365738] env[63372]: DEBUG nova.network.neutron [req-d31e204b-df4f-46f8-9a91-12148e57afca req-3dc55255-e455-437a-83d2-c2f47e191444 service nova] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.382661] env[63372]: DEBUG nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 531.415967] env[63372]: DEBUG nova.virt.hardware [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 531.416257] env[63372]: DEBUG nova.virt.hardware [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 531.416433] env[63372]: DEBUG nova.virt.hardware [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.416662] env[63372]: DEBUG nova.virt.hardware [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 531.420138] env[63372]: DEBUG nova.virt.hardware [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.420348] env[63372]: DEBUG nova.virt.hardware [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 531.420567] env[63372]: DEBUG nova.virt.hardware [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 531.420729] env[63372]: DEBUG nova.virt.hardware [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 531.420895] env[63372]: DEBUG nova.virt.hardware [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 531.421067] env[63372]: DEBUG nova.virt.hardware [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 531.421740] env[63372]: DEBUG nova.virt.hardware [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 531.422128] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2f4c85-461b-4de0-ab11-97a9c5af1350 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.425036] env[63372]: DEBUG nova.network.neutron [-] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.431377] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dbe2d9d-05bf-41f5-b62c-f97c6c7d470d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.585493] env[63372]: DEBUG nova.network.neutron [req-d31e204b-df4f-46f8-9a91-12148e57afca req-3dc55255-e455-437a-83d2-c2f47e191444 service nova] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.593172] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.238s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.593172] env[63372]: DEBUG nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 531.599197] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.030s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.606995] env[63372]: INFO nova.compute.claims [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 531.651830] env[63372]: DEBUG nova.network.neutron [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Successfully created port: 8c633cc9-a20a-4af4-ad1f-b367ad5c9fa9 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 531.929667] env[63372]: INFO nova.compute.manager [-] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Took 1.06 seconds to deallocate network for instance. [ 531.934460] env[63372]: DEBUG nova.compute.claims [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 531.934460] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.088868] env[63372]: DEBUG oslo_concurrency.lockutils [req-d31e204b-df4f-46f8-9a91-12148e57afca req-3dc55255-e455-437a-83d2-c2f47e191444 service nova] Releasing lock "refresh_cache-5cf85718-bb73-4534-9dd9-ddb4b6777f6e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.102212] env[63372]: DEBUG nova.compute.utils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.102212] env[63372]: DEBUG nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 532.102212] env[63372]: DEBUG nova.network.neutron [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 532.251322] env[63372]: DEBUG nova.policy [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f8e364e3aed40248cc0d11741f76154', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b4548b52cd704cc0b054e2f7d1562b9a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 532.610209] env[63372]: DEBUG nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 532.756183] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4df6d0a-4479-4573-a05c-1bf5de0b7c23 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.767014] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f893d1ef-0d5b-425b-96a6-2e9b37eb5dd1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.800841] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51f5068c-e5ad-447b-88c0-36fd9cbb7dd1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.809036] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f0749e7-45e7-4b4d-9a75-d6b67915125a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.824619] env[63372]: DEBUG nova.compute.provider_tree [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 533.320913] env[63372]: ERROR nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6590d44b-a55c-49b4-8105-3915d4a6d5a2, please check neutron logs for more information. [ 533.320913] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 533.320913] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.320913] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 533.320913] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.320913] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 533.320913] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.320913] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 533.320913] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.320913] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 533.320913] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.320913] env[63372]: ERROR nova.compute.manager raise self.value [ 533.320913] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.320913] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 533.320913] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.320913] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 533.321329] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.321329] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 533.321329] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6590d44b-a55c-49b4-8105-3915d4a6d5a2, please check neutron logs for more information. [ 533.321329] env[63372]: ERROR nova.compute.manager [ 533.321329] env[63372]: Traceback (most recent call last): [ 533.321329] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 533.321329] env[63372]: listener.cb(fileno) [ 533.321329] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.321329] env[63372]: result = function(*args, **kwargs) [ 533.321329] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.321329] env[63372]: return func(*args, **kwargs) [ 533.321329] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.321329] env[63372]: raise e [ 533.321329] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.321329] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 533.321329] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.321329] env[63372]: created_port_ids = self._update_ports_for_instance( [ 533.321329] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.321329] env[63372]: with excutils.save_and_reraise_exception(): [ 533.321329] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.321329] env[63372]: self.force_reraise() [ 533.321329] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.321329] env[63372]: raise self.value [ 533.321329] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.321329] env[63372]: updated_port = self._update_port( [ 533.321329] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.321329] env[63372]: _ensure_no_port_binding_failure(port) [ 533.321329] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.321329] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 533.322058] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 6590d44b-a55c-49b4-8105-3915d4a6d5a2, please check neutron logs for more information. [ 533.322058] env[63372]: Removing descriptor: 16 [ 533.322058] env[63372]: ERROR nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6590d44b-a55c-49b4-8105-3915d4a6d5a2, please check neutron logs for more information. [ 533.322058] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Traceback (most recent call last): [ 533.322058] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 533.322058] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] yield resources [ 533.322058] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 533.322058] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] self.driver.spawn(context, instance, image_meta, [ 533.322058] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 533.322058] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.322058] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.322058] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] vm_ref = self.build_virtual_machine(instance, [ 533.322441] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.322441] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.322441] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.322441] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] for vif in network_info: [ 533.322441] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 533.322441] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] return self._sync_wrapper(fn, *args, **kwargs) [ 533.322441] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 533.322441] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] self.wait() [ 533.322441] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 533.322441] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] self[:] = self._gt.wait() [ 533.322441] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.322441] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] return self._exit_event.wait() [ 533.322441] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.322782] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] result = hub.switch() [ 533.322782] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.322782] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] return self.greenlet.switch() [ 533.322782] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.322782] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] result = function(*args, **kwargs) [ 533.322782] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.322782] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] return func(*args, **kwargs) [ 533.322782] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.322782] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] raise e [ 533.322782] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.322782] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] nwinfo = self.network_api.allocate_for_instance( [ 533.322782] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.322782] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] created_port_ids = self._update_ports_for_instance( [ 533.323128] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.323128] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] with excutils.save_and_reraise_exception(): [ 533.323128] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.323128] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] self.force_reraise() [ 533.323128] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.323128] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] raise self.value [ 533.323128] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.323128] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] updated_port = self._update_port( [ 533.323128] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.323128] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] _ensure_no_port_binding_failure(port) [ 533.323128] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.323128] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] raise exception.PortBindingFailed(port_id=port['id']) [ 533.323451] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] nova.exception.PortBindingFailed: Binding failed for port 6590d44b-a55c-49b4-8105-3915d4a6d5a2, please check neutron logs for more information. [ 533.323451] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] [ 533.323451] env[63372]: INFO nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Terminating instance [ 533.326129] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Acquiring lock "refresh_cache-f33eea3e-887c-4164-98e9-e4d502d14af2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.326129] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Acquired lock "refresh_cache-f33eea3e-887c-4164-98e9-e4d502d14af2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.326129] env[63372]: DEBUG nova.network.neutron [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 533.326433] env[63372]: DEBUG nova.scheduler.client.report [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 533.620479] env[63372]: DEBUG nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 533.656359] env[63372]: DEBUG nova.virt.hardware [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 533.656885] env[63372]: DEBUG nova.virt.hardware [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 533.657222] env[63372]: DEBUG nova.virt.hardware [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 533.657541] env[63372]: DEBUG nova.virt.hardware [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 533.659073] env[63372]: DEBUG nova.virt.hardware [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 533.659073] env[63372]: DEBUG nova.virt.hardware [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 533.659073] env[63372]: DEBUG nova.virt.hardware [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 533.659073] env[63372]: DEBUG nova.virt.hardware [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 533.659073] env[63372]: DEBUG nova.virt.hardware [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 533.659276] env[63372]: DEBUG nova.virt.hardware [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 533.659276] env[63372]: DEBUG nova.virt.hardware [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 533.663320] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ea72c8-7079-4a74-b4ef-53eeafae375b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.669657] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f94df9b-918f-4566-becf-6064fc075e57 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.701083] env[63372]: DEBUG nova.network.neutron [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Successfully created port: c92db39d-fd16-4e13-a151-e5781476ef44 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 533.835468] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.235s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.835468] env[63372]: DEBUG nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 533.838124] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 1.904s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.891671] env[63372]: DEBUG nova.network.neutron [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.053478] env[63372]: ERROR nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5f466550-60f4-48bb-871d-060148bd4558, please check neutron logs for more information. [ 534.053478] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 534.053478] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.053478] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 534.053478] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.053478] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 534.053478] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.053478] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 534.053478] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.053478] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 534.053478] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.053478] env[63372]: ERROR nova.compute.manager raise self.value [ 534.053478] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.053478] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 534.053478] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.053478] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 534.053930] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.053930] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 534.053930] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5f466550-60f4-48bb-871d-060148bd4558, please check neutron logs for more information. [ 534.053930] env[63372]: ERROR nova.compute.manager [ 534.053930] env[63372]: Traceback (most recent call last): [ 534.053930] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 534.053930] env[63372]: listener.cb(fileno) [ 534.053930] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.053930] env[63372]: result = function(*args, **kwargs) [ 534.053930] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 534.053930] env[63372]: return func(*args, **kwargs) [ 534.053930] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.053930] env[63372]: raise e [ 534.053930] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.053930] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 534.053930] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.053930] env[63372]: created_port_ids = self._update_ports_for_instance( [ 534.053930] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.053930] env[63372]: with excutils.save_and_reraise_exception(): [ 534.053930] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.053930] env[63372]: self.force_reraise() [ 534.053930] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.053930] env[63372]: raise self.value [ 534.053930] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.053930] env[63372]: updated_port = self._update_port( [ 534.053930] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.053930] env[63372]: _ensure_no_port_binding_failure(port) [ 534.053930] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.053930] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 534.054630] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 5f466550-60f4-48bb-871d-060148bd4558, please check neutron logs for more information. [ 534.054630] env[63372]: Removing descriptor: 17 [ 534.054630] env[63372]: ERROR nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5f466550-60f4-48bb-871d-060148bd4558, please check neutron logs for more information. [ 534.054630] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Traceback (most recent call last): [ 534.054630] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 534.054630] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] yield resources [ 534.054630] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 534.054630] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] self.driver.spawn(context, instance, image_meta, [ 534.054630] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 534.054630] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] self._vmops.spawn(context, instance, image_meta, injected_files, [ 534.054630] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 534.054630] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] vm_ref = self.build_virtual_machine(instance, [ 534.054983] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 534.054983] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] vif_infos = vmwarevif.get_vif_info(self._session, [ 534.054983] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 534.054983] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] for vif in network_info: [ 534.054983] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 534.054983] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] return self._sync_wrapper(fn, *args, **kwargs) [ 534.054983] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 534.054983] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] self.wait() [ 534.054983] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 534.054983] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] self[:] = self._gt.wait() [ 534.054983] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 534.054983] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] return self._exit_event.wait() [ 534.054983] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 534.055569] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] result = hub.switch() [ 534.055569] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 534.055569] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] return self.greenlet.switch() [ 534.055569] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.055569] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] result = function(*args, **kwargs) [ 534.055569] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 534.055569] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] return func(*args, **kwargs) [ 534.055569] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.055569] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] raise e [ 534.055569] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.055569] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] nwinfo = self.network_api.allocate_for_instance( [ 534.055569] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.055569] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] created_port_ids = self._update_ports_for_instance( [ 534.056486] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.056486] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] with excutils.save_and_reraise_exception(): [ 534.056486] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.056486] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] self.force_reraise() [ 534.056486] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.056486] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] raise self.value [ 534.056486] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.056486] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] updated_port = self._update_port( [ 534.056486] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.056486] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] _ensure_no_port_binding_failure(port) [ 534.056486] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.056486] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] raise exception.PortBindingFailed(port_id=port['id']) [ 534.056808] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] nova.exception.PortBindingFailed: Binding failed for port 5f466550-60f4-48bb-871d-060148bd4558, please check neutron logs for more information. [ 534.056808] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] [ 534.056808] env[63372]: INFO nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Terminating instance [ 534.059828] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Acquiring lock "refresh_cache-393f3218-5e22-4e88-8968-9157aa0631da" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.060240] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Acquired lock "refresh_cache-393f3218-5e22-4e88-8968-9157aa0631da" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.060308] env[63372]: DEBUG nova.network.neutron [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 534.066537] env[63372]: DEBUG nova.network.neutron [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.137273] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.137561] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.137754] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Starting heal instance info cache {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 534.137900] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Rebuilding the list of instances to heal {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 534.342854] env[63372]: DEBUG nova.compute.utils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 534.347271] env[63372]: DEBUG nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 534.348942] env[63372]: DEBUG nova.network.neutron [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 534.456597] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6b694a-9f12-4026-b6d1-6f15f6cf3a7f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.468140] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466560d6-c356-452a-ad16-8cdcf95e0371 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.502031] env[63372]: DEBUG nova.policy [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cfb8b4e8f7f9482da29a00921083acff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c57ad37f4ae3462e9a2cecd6fcb9919a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 534.504716] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-503a8c7f-c986-4353-b31c-af01fc2f9575 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.512605] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b7cd4bc-ba27-4834-8159-68c10e32226b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.529459] env[63372]: DEBUG nova.compute.provider_tree [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 534.567391] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Releasing lock "refresh_cache-f33eea3e-887c-4164-98e9-e4d502d14af2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.567955] env[63372]: DEBUG nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 534.568424] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 534.569108] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-082058b2-0d41-4223-a877-174b85938fa9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.581340] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80369774-935b-42e8-a170-2bbcb2d4802e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.603551] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f33eea3e-887c-4164-98e9-e4d502d14af2 could not be found. [ 534.604317] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 534.604361] env[63372]: INFO nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 534.604710] env[63372]: DEBUG oslo.service.loopingcall [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 534.604821] env[63372]: DEBUG nova.compute.manager [-] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 534.604901] env[63372]: DEBUG nova.network.neutron [-] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 534.622824] env[63372]: DEBUG nova.network.neutron [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.646492] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 534.646807] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 534.647042] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 32dabedb-8155-4377-871b-408935db97b5] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 534.647316] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 534.647530] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 534.648029] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Didn't find any instances for network info cache update. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 534.648029] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.648260] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.648448] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.649424] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.649424] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.649424] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.649424] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63372) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 534.649424] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager.update_available_resource {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.672285] env[63372]: DEBUG nova.network.neutron [-] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.850647] env[63372]: DEBUG nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 534.890086] env[63372]: DEBUG nova.network.neutron [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.034591] env[63372]: DEBUG nova.scheduler.client.report [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 535.156814] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.173437] env[63372]: DEBUG nova.network.neutron [-] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.213214] env[63372]: DEBUG nova.compute.manager [req-6f7370fb-3eb1-41fb-aef9-fe6e9ebbdea8 req-13d5cdad-9de8-4a67-9f0b-1f465ec132bb service nova] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Received event network-changed-6590d44b-a55c-49b4-8105-3915d4a6d5a2 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 535.214043] env[63372]: DEBUG nova.compute.manager [req-6f7370fb-3eb1-41fb-aef9-fe6e9ebbdea8 req-13d5cdad-9de8-4a67-9f0b-1f465ec132bb service nova] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Refreshing instance network info cache due to event network-changed-6590d44b-a55c-49b4-8105-3915d4a6d5a2. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 535.214043] env[63372]: DEBUG oslo_concurrency.lockutils [req-6f7370fb-3eb1-41fb-aef9-fe6e9ebbdea8 req-13d5cdad-9de8-4a67-9f0b-1f465ec132bb service nova] Acquiring lock "refresh_cache-f33eea3e-887c-4164-98e9-e4d502d14af2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.214043] env[63372]: DEBUG oslo_concurrency.lockutils [req-6f7370fb-3eb1-41fb-aef9-fe6e9ebbdea8 req-13d5cdad-9de8-4a67-9f0b-1f465ec132bb service nova] Acquired lock "refresh_cache-f33eea3e-887c-4164-98e9-e4d502d14af2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.214043] env[63372]: DEBUG nova.network.neutron [req-6f7370fb-3eb1-41fb-aef9-fe6e9ebbdea8 req-13d5cdad-9de8-4a67-9f0b-1f465ec132bb service nova] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Refreshing network info cache for port 6590d44b-a55c-49b4-8105-3915d4a6d5a2 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 535.397114] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Releasing lock "refresh_cache-393f3218-5e22-4e88-8968-9157aa0631da" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.397114] env[63372]: DEBUG nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 535.397114] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 535.397114] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51966f53-66d1-4cca-9ea9-34576dcdc1db {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.415998] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686dab5e-5e12-4d45-95de-761da111f764 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.444891] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 393f3218-5e22-4e88-8968-9157aa0631da could not be found. [ 535.448277] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 535.448693] env[63372]: INFO nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Took 0.06 seconds to destroy the instance on the hypervisor. [ 535.448693] env[63372]: DEBUG oslo.service.loopingcall [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.449438] env[63372]: DEBUG nova.compute.manager [-] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.449438] env[63372]: DEBUG nova.network.neutron [-] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 535.503817] env[63372]: DEBUG nova.network.neutron [-] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.541858] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.704s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.545188] env[63372]: ERROR nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8c7ae2fc-56c6-43bd-8f50-5a1eebe4fa92, please check neutron logs for more information. [ 535.545188] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Traceback (most recent call last): [ 535.545188] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 535.545188] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] self.driver.spawn(context, instance, image_meta, [ 535.545188] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 535.545188] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 535.545188] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 535.545188] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] vm_ref = self.build_virtual_machine(instance, [ 535.545188] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 535.545188] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] vif_infos = vmwarevif.get_vif_info(self._session, [ 535.545188] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 535.545655] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] for vif in network_info: [ 535.545655] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 535.545655] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] return self._sync_wrapper(fn, *args, **kwargs) [ 535.545655] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 535.545655] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] self.wait() [ 535.545655] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 535.545655] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] self[:] = self._gt.wait() [ 535.545655] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 535.545655] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] return self._exit_event.wait() [ 535.545655] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 535.545655] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] result = hub.switch() [ 535.545655] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 535.545655] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] return self.greenlet.switch() [ 535.545979] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.545979] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] result = function(*args, **kwargs) [ 535.545979] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.545979] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] return func(*args, **kwargs) [ 535.545979] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.545979] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] raise e [ 535.545979] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.545979] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] nwinfo = self.network_api.allocate_for_instance( [ 535.545979] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.545979] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] created_port_ids = self._update_ports_for_instance( [ 535.545979] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.545979] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] with excutils.save_and_reraise_exception(): [ 535.545979] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.546307] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] self.force_reraise() [ 535.546307] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.546307] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] raise self.value [ 535.546307] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.546307] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] updated_port = self._update_port( [ 535.546307] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.546307] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] _ensure_no_port_binding_failure(port) [ 535.546307] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.546307] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] raise exception.PortBindingFailed(port_id=port['id']) [ 535.546307] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] nova.exception.PortBindingFailed: Binding failed for port 8c7ae2fc-56c6-43bd-8f50-5a1eebe4fa92, please check neutron logs for more information. [ 535.546307] env[63372]: ERROR nova.compute.manager [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] [ 535.546573] env[63372]: DEBUG nova.compute.utils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Binding failed for port 8c7ae2fc-56c6-43bd-8f50-5a1eebe4fa92, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 535.546573] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.389s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.546573] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.546573] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63372) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 535.546573] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f914da74-4e8c-4f27-bc1f-295641843bbe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.552935] env[63372]: DEBUG nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Build of instance 5cf85718-bb73-4534-9dd9-ddb4b6777f6e was re-scheduled: Binding failed for port 8c7ae2fc-56c6-43bd-8f50-5a1eebe4fa92, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 535.552935] env[63372]: DEBUG nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 535.552935] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Acquiring lock "refresh_cache-5cf85718-bb73-4534-9dd9-ddb4b6777f6e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.552935] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Acquired lock "refresh_cache-5cf85718-bb73-4534-9dd9-ddb4b6777f6e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.553365] env[63372]: DEBUG nova.network.neutron [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 535.560929] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33aca790-2bcc-4318-a7a5-08184fd318c3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.584321] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29dab705-52aa-4f63-bf5a-dcc14c40b63f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.594982] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-822e909b-bba3-4c10-8b2b-a8aee5e26606 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.629329] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181576MB free_disk=186GB free_vcpus=48 pci_devices=None {{(pid=63372) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 535.629590] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.630754] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.676672] env[63372]: INFO nova.compute.manager [-] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Took 1.07 seconds to deallocate network for instance. [ 535.680355] env[63372]: DEBUG nova.compute.claims [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 535.680669] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.782116] env[63372]: DEBUG nova.network.neutron [req-6f7370fb-3eb1-41fb-aef9-fe6e9ebbdea8 req-13d5cdad-9de8-4a67-9f0b-1f465ec132bb service nova] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.868183] env[63372]: DEBUG nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 535.896912] env[63372]: DEBUG nova.virt.hardware [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 535.897152] env[63372]: DEBUG nova.virt.hardware [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 535.897304] env[63372]: DEBUG nova.virt.hardware [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 535.897486] env[63372]: DEBUG nova.virt.hardware [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 535.897610] env[63372]: DEBUG nova.virt.hardware [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 535.897780] env[63372]: DEBUG nova.virt.hardware [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 535.897994] env[63372]: DEBUG nova.virt.hardware [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 535.898160] env[63372]: DEBUG nova.virt.hardware [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 535.898315] env[63372]: DEBUG nova.virt.hardware [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 535.898503] env[63372]: DEBUG nova.virt.hardware [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 535.898624] env[63372]: DEBUG nova.virt.hardware [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 535.899587] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57806f3d-adbf-44be-b0d7-1d6f4967711f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.907673] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-025768b9-8acc-471c-a2c3-a2494ff40ef7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.007601] env[63372]: DEBUG nova.network.neutron [-] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.067405] env[63372]: DEBUG nova.network.neutron [req-6f7370fb-3eb1-41fb-aef9-fe6e9ebbdea8 req-13d5cdad-9de8-4a67-9f0b-1f465ec132bb service nova] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.120639] env[63372]: DEBUG nova.network.neutron [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.364806] env[63372]: DEBUG nova.network.neutron [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.471132] env[63372]: DEBUG nova.network.neutron [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Successfully created port: 69372e7a-1cad-4678-8591-897c91ccc7e1 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 536.511232] env[63372]: ERROR nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8c633cc9-a20a-4af4-ad1f-b367ad5c9fa9, please check neutron logs for more information. [ 536.511232] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 536.511232] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.511232] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 536.511232] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.511232] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 536.511232] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.511232] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 536.511232] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.511232] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 536.511232] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.511232] env[63372]: ERROR nova.compute.manager raise self.value [ 536.511232] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.511232] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 536.511232] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.511232] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 536.511739] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.511739] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 536.511739] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8c633cc9-a20a-4af4-ad1f-b367ad5c9fa9, please check neutron logs for more information. [ 536.511739] env[63372]: ERROR nova.compute.manager [ 536.511739] env[63372]: Traceback (most recent call last): [ 536.511739] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 536.511739] env[63372]: listener.cb(fileno) [ 536.511739] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.511739] env[63372]: result = function(*args, **kwargs) [ 536.511739] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.511739] env[63372]: return func(*args, **kwargs) [ 536.511739] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.511739] env[63372]: raise e [ 536.511739] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.511739] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 536.511739] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.511739] env[63372]: created_port_ids = self._update_ports_for_instance( [ 536.511739] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.511739] env[63372]: with excutils.save_and_reraise_exception(): [ 536.511739] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.511739] env[63372]: self.force_reraise() [ 536.511739] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.511739] env[63372]: raise self.value [ 536.511739] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.511739] env[63372]: updated_port = self._update_port( [ 536.511739] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.511739] env[63372]: _ensure_no_port_binding_failure(port) [ 536.511739] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.511739] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 536.512645] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 8c633cc9-a20a-4af4-ad1f-b367ad5c9fa9, please check neutron logs for more information. [ 536.512645] env[63372]: Removing descriptor: 18 [ 536.512645] env[63372]: INFO nova.compute.manager [-] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Took 1.06 seconds to deallocate network for instance. [ 536.512889] env[63372]: ERROR nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8c633cc9-a20a-4af4-ad1f-b367ad5c9fa9, please check neutron logs for more information. [ 536.512889] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] Traceback (most recent call last): [ 536.512889] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 536.512889] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] yield resources [ 536.512889] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 536.512889] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] self.driver.spawn(context, instance, image_meta, [ 536.512889] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 536.512889] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.512889] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.512889] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] vm_ref = self.build_virtual_machine(instance, [ 536.512889] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.513252] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.513252] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.513252] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] for vif in network_info: [ 536.513252] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 536.513252] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] return self._sync_wrapper(fn, *args, **kwargs) [ 536.513252] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 536.513252] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] self.wait() [ 536.513252] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 536.513252] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] self[:] = self._gt.wait() [ 536.513252] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.513252] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] return self._exit_event.wait() [ 536.513252] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.513252] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] result = hub.switch() [ 536.513859] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.513859] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] return self.greenlet.switch() [ 536.513859] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.513859] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] result = function(*args, **kwargs) [ 536.513859] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.513859] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] return func(*args, **kwargs) [ 536.513859] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.513859] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] raise e [ 536.513859] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.513859] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] nwinfo = self.network_api.allocate_for_instance( [ 536.513859] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.513859] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] created_port_ids = self._update_ports_for_instance( [ 536.513859] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.514423] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] with excutils.save_and_reraise_exception(): [ 536.514423] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.514423] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] self.force_reraise() [ 536.514423] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.514423] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] raise self.value [ 536.514423] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.514423] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] updated_port = self._update_port( [ 536.514423] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.514423] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] _ensure_no_port_binding_failure(port) [ 536.514423] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.514423] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] raise exception.PortBindingFailed(port_id=port['id']) [ 536.514423] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] nova.exception.PortBindingFailed: Binding failed for port 8c633cc9-a20a-4af4-ad1f-b367ad5c9fa9, please check neutron logs for more information. [ 536.514423] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] [ 536.514796] env[63372]: INFO nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Terminating instance [ 536.520184] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Acquiring lock "refresh_cache-32dabedb-8155-4377-871b-408935db97b5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.520184] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Acquired lock "refresh_cache-32dabedb-8155-4377-871b-408935db97b5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.520184] env[63372]: DEBUG nova.network.neutron [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 536.529822] env[63372]: DEBUG nova.compute.claims [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 536.530018] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.570932] env[63372]: DEBUG oslo_concurrency.lockutils [req-6f7370fb-3eb1-41fb-aef9-fe6e9ebbdea8 req-13d5cdad-9de8-4a67-9f0b-1f465ec132bb service nova] Releasing lock "refresh_cache-f33eea3e-887c-4164-98e9-e4d502d14af2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.867388] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Releasing lock "refresh_cache-5cf85718-bb73-4534-9dd9-ddb4b6777f6e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.867622] env[63372]: DEBUG nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 536.867774] env[63372]: DEBUG nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 536.867928] env[63372]: DEBUG nova.network.neutron [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 536.929577] env[63372]: DEBUG nova.network.neutron [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.091487] env[63372]: DEBUG nova.network.neutron [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.175580] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 5cf85718-bb73-4534-9dd9-ddb4b6777f6e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 537.176029] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance f33eea3e-887c-4164-98e9-e4d502d14af2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.176029] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 393f3218-5e22-4e88-8968-9157aa0631da actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.177122] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 32dabedb-8155-4377-871b-408935db97b5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.177122] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 21c4415c-a016-47eb-9ec4-280c6c18b3fb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.177122] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 91351ef7-a282-4f57-b96b-ba0af6d0e074 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.177122] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 537.177122] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 537.340151] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc3e3f2-4755-44ba-aba0-2e5339f16b09 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.349136] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1659786-3935-4312-bd86-f7eef2276843 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.386347] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfdc3a41-898a-4279-a6f9-c93236cf2e22 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.393361] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fa7402-c2c7-44eb-a3e3-304d124c8379 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.408933] env[63372]: DEBUG nova.compute.provider_tree [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 537.436059] env[63372]: DEBUG nova.network.neutron [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.538347] env[63372]: DEBUG nova.network.neutron [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.733900] env[63372]: ERROR nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c92db39d-fd16-4e13-a151-e5781476ef44, please check neutron logs for more information. [ 537.733900] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 537.733900] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.733900] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 537.733900] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.733900] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 537.733900] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.733900] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 537.733900] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.733900] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 537.733900] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.733900] env[63372]: ERROR nova.compute.manager raise self.value [ 537.733900] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.733900] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 537.733900] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.733900] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 537.734679] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.734679] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 537.734679] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c92db39d-fd16-4e13-a151-e5781476ef44, please check neutron logs for more information. [ 537.734679] env[63372]: ERROR nova.compute.manager [ 537.734679] env[63372]: Traceback (most recent call last): [ 537.734679] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 537.734679] env[63372]: listener.cb(fileno) [ 537.734679] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.734679] env[63372]: result = function(*args, **kwargs) [ 537.734679] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.734679] env[63372]: return func(*args, **kwargs) [ 537.734679] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.734679] env[63372]: raise e [ 537.734679] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.734679] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 537.734679] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.734679] env[63372]: created_port_ids = self._update_ports_for_instance( [ 537.734679] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.734679] env[63372]: with excutils.save_and_reraise_exception(): [ 537.734679] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.734679] env[63372]: self.force_reraise() [ 537.734679] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.734679] env[63372]: raise self.value [ 537.734679] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.734679] env[63372]: updated_port = self._update_port( [ 537.734679] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.734679] env[63372]: _ensure_no_port_binding_failure(port) [ 537.734679] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.734679] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 537.735388] env[63372]: nova.exception.PortBindingFailed: Binding failed for port c92db39d-fd16-4e13-a151-e5781476ef44, please check neutron logs for more information. [ 537.735388] env[63372]: Removing descriptor: 14 [ 537.735388] env[63372]: ERROR nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c92db39d-fd16-4e13-a151-e5781476ef44, please check neutron logs for more information. [ 537.735388] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Traceback (most recent call last): [ 537.735388] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 537.735388] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] yield resources [ 537.735388] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 537.735388] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] self.driver.spawn(context, instance, image_meta, [ 537.735388] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 537.735388] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.735388] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.735388] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] vm_ref = self.build_virtual_machine(instance, [ 537.735692] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.735692] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.735692] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.735692] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] for vif in network_info: [ 537.735692] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 537.735692] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] return self._sync_wrapper(fn, *args, **kwargs) [ 537.735692] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 537.735692] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] self.wait() [ 537.735692] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 537.735692] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] self[:] = self._gt.wait() [ 537.735692] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.735692] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] return self._exit_event.wait() [ 537.735692] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.736017] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] result = hub.switch() [ 537.736017] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.736017] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] return self.greenlet.switch() [ 537.736017] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.736017] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] result = function(*args, **kwargs) [ 537.736017] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.736017] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] return func(*args, **kwargs) [ 537.736017] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.736017] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] raise e [ 537.736017] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.736017] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] nwinfo = self.network_api.allocate_for_instance( [ 537.736017] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.736017] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] created_port_ids = self._update_ports_for_instance( [ 537.736341] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.736341] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] with excutils.save_and_reraise_exception(): [ 537.736341] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.736341] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] self.force_reraise() [ 537.736341] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.736341] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] raise self.value [ 537.736341] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.736341] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] updated_port = self._update_port( [ 537.736341] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.736341] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] _ensure_no_port_binding_failure(port) [ 537.736341] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.736341] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] raise exception.PortBindingFailed(port_id=port['id']) [ 537.737790] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] nova.exception.PortBindingFailed: Binding failed for port c92db39d-fd16-4e13-a151-e5781476ef44, please check neutron logs for more information. [ 537.737790] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] [ 537.737790] env[63372]: INFO nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Terminating instance [ 537.737790] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "refresh_cache-21c4415c-a016-47eb-9ec4-280c6c18b3fb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.737790] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired lock "refresh_cache-21c4415c-a016-47eb-9ec4-280c6c18b3fb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.738143] env[63372]: DEBUG nova.network.neutron [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 537.911976] env[63372]: DEBUG nova.scheduler.client.report [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 537.938673] env[63372]: INFO nova.compute.manager [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 5cf85718-bb73-4534-9dd9-ddb4b6777f6e] Took 1.07 seconds to deallocate network for instance. [ 538.045117] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Releasing lock "refresh_cache-32dabedb-8155-4377-871b-408935db97b5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.045117] env[63372]: DEBUG nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 538.045117] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 538.045117] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-90ff6878-3851-4876-bc0d-7bfbf6edf8b3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.056610] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d91c301-c8fc-4a22-9eed-d938eacf63a0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.082376] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 32dabedb-8155-4377-871b-408935db97b5 could not be found. [ 538.082688] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 538.082891] env[63372]: INFO nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 538.083208] env[63372]: DEBUG oslo.service.loopingcall [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 538.083519] env[63372]: DEBUG nova.compute.manager [-] [instance: 32dabedb-8155-4377-871b-408935db97b5] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.083519] env[63372]: DEBUG nova.network.neutron [-] [instance: 32dabedb-8155-4377-871b-408935db97b5] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 538.219423] env[63372]: DEBUG nova.network.neutron [-] [instance: 32dabedb-8155-4377-871b-408935db97b5] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.264247] env[63372]: DEBUG nova.network.neutron [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.407274] env[63372]: DEBUG nova.network.neutron [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.417221] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 538.417796] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.788s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.418215] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 2.738s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.726684] env[63372]: DEBUG nova.network.neutron [-] [instance: 32dabedb-8155-4377-871b-408935db97b5] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.909486] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Releasing lock "refresh_cache-21c4415c-a016-47eb-9ec4-280c6c18b3fb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.909974] env[63372]: DEBUG nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 538.910198] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 538.910502] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-068deeab-27d4-46c6-b55a-b6f609bb9d53 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.927174] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2758dc76-a497-4364-b2de-6ddf104e65e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.957503] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 21c4415c-a016-47eb-9ec4-280c6c18b3fb could not be found. [ 538.957720] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 538.957895] env[63372]: INFO nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Took 0.05 seconds to destroy the instance on the hypervisor. [ 538.958400] env[63372]: DEBUG oslo.service.loopingcall [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 538.958954] env[63372]: DEBUG nova.compute.manager [-] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.959076] env[63372]: DEBUG nova.network.neutron [-] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 538.976768] env[63372]: INFO nova.scheduler.client.report [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Deleted allocations for instance 5cf85718-bb73-4534-9dd9-ddb4b6777f6e [ 538.984781] env[63372]: DEBUG nova.network.neutron [-] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.075904] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3c975f-88bb-4463-9534-4ad6a6ddae3d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.084828] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b48e5f1-8e1b-4205-b3bf-5a5dd413bb9d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.117052] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-969a14f1-63d6-467f-8dc3-8812238c017e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.128643] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fda5624-0e6d-4cea-a63e-70f005cf2632 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.146654] env[63372]: DEBUG nova.compute.provider_tree [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 539.229490] env[63372]: INFO nova.compute.manager [-] [instance: 32dabedb-8155-4377-871b-408935db97b5] Took 1.15 seconds to deallocate network for instance. [ 539.233687] env[63372]: DEBUG nova.compute.claims [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 539.233796] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.308220] env[63372]: DEBUG nova.compute.manager [req-fe5832f0-73d0-4276-8ca7-c93ea7561df9 req-0a2d85ea-accc-4ddc-8295-6abad0d4752c service nova] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Received event network-changed-5f466550-60f4-48bb-871d-060148bd4558 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.308220] env[63372]: DEBUG nova.compute.manager [req-fe5832f0-73d0-4276-8ca7-c93ea7561df9 req-0a2d85ea-accc-4ddc-8295-6abad0d4752c service nova] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Refreshing instance network info cache due to event network-changed-5f466550-60f4-48bb-871d-060148bd4558. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 539.308220] env[63372]: DEBUG oslo_concurrency.lockutils [req-fe5832f0-73d0-4276-8ca7-c93ea7561df9 req-0a2d85ea-accc-4ddc-8295-6abad0d4752c service nova] Acquiring lock "refresh_cache-393f3218-5e22-4e88-8968-9157aa0631da" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.308220] env[63372]: DEBUG oslo_concurrency.lockutils [req-fe5832f0-73d0-4276-8ca7-c93ea7561df9 req-0a2d85ea-accc-4ddc-8295-6abad0d4752c service nova] Acquired lock "refresh_cache-393f3218-5e22-4e88-8968-9157aa0631da" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.308220] env[63372]: DEBUG nova.network.neutron [req-fe5832f0-73d0-4276-8ca7-c93ea7561df9 req-0a2d85ea-accc-4ddc-8295-6abad0d4752c service nova] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Refreshing network info cache for port 5f466550-60f4-48bb-871d-060148bd4558 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 539.335846] env[63372]: DEBUG nova.compute.manager [req-ba4fbfb6-c299-45e0-8c40-caf648d15146 req-d6fd5117-fd6b-47c3-a8a5-3b71fc113bb2 service nova] [instance: 32dabedb-8155-4377-871b-408935db97b5] Received event network-changed-8c633cc9-a20a-4af4-ad1f-b367ad5c9fa9 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.337075] env[63372]: DEBUG nova.compute.manager [req-ba4fbfb6-c299-45e0-8c40-caf648d15146 req-d6fd5117-fd6b-47c3-a8a5-3b71fc113bb2 service nova] [instance: 32dabedb-8155-4377-871b-408935db97b5] Refreshing instance network info cache due to event network-changed-8c633cc9-a20a-4af4-ad1f-b367ad5c9fa9. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 539.337415] env[63372]: DEBUG oslo_concurrency.lockutils [req-ba4fbfb6-c299-45e0-8c40-caf648d15146 req-d6fd5117-fd6b-47c3-a8a5-3b71fc113bb2 service nova] Acquiring lock "refresh_cache-32dabedb-8155-4377-871b-408935db97b5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.337563] env[63372]: DEBUG oslo_concurrency.lockutils [req-ba4fbfb6-c299-45e0-8c40-caf648d15146 req-d6fd5117-fd6b-47c3-a8a5-3b71fc113bb2 service nova] Acquired lock "refresh_cache-32dabedb-8155-4377-871b-408935db97b5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.337723] env[63372]: DEBUG nova.network.neutron [req-ba4fbfb6-c299-45e0-8c40-caf648d15146 req-d6fd5117-fd6b-47c3-a8a5-3b71fc113bb2 service nova] [instance: 32dabedb-8155-4377-871b-408935db97b5] Refreshing network info cache for port 8c633cc9-a20a-4af4-ad1f-b367ad5c9fa9 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 539.350212] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Acquiring lock "d290c8b5-422b-43c6-825e-52ea98516e6c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.351057] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Lock "d290c8b5-422b-43c6-825e-52ea98516e6c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.487027] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1d3b419a-0bd9-4014-adeb-bb2b5ce2dfee tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Lock "5cf85718-bb73-4534-9dd9-ddb4b6777f6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.957s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.490137] env[63372]: DEBUG nova.network.neutron [-] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.650469] env[63372]: DEBUG nova.scheduler.client.report [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 539.848033] env[63372]: DEBUG nova.network.neutron [req-fe5832f0-73d0-4276-8ca7-c93ea7561df9 req-0a2d85ea-accc-4ddc-8295-6abad0d4752c service nova] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.860030] env[63372]: DEBUG nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 539.899493] env[63372]: DEBUG nova.network.neutron [req-ba4fbfb6-c299-45e0-8c40-caf648d15146 req-d6fd5117-fd6b-47c3-a8a5-3b71fc113bb2 service nova] [instance: 32dabedb-8155-4377-871b-408935db97b5] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.923406] env[63372]: DEBUG nova.network.neutron [req-fe5832f0-73d0-4276-8ca7-c93ea7561df9 req-0a2d85ea-accc-4ddc-8295-6abad0d4752c service nova] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.994091] env[63372]: INFO nova.compute.manager [-] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Took 1.03 seconds to deallocate network for instance. [ 539.999565] env[63372]: DEBUG nova.compute.claims [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 539.999825] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.162815] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.744s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.163580] env[63372]: ERROR nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6590d44b-a55c-49b4-8105-3915d4a6d5a2, please check neutron logs for more information. [ 540.163580] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Traceback (most recent call last): [ 540.163580] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 540.163580] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] self.driver.spawn(context, instance, image_meta, [ 540.163580] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 540.163580] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.163580] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.163580] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] vm_ref = self.build_virtual_machine(instance, [ 540.163580] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.163580] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.163580] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.166987] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] for vif in network_info: [ 540.166987] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.166987] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] return self._sync_wrapper(fn, *args, **kwargs) [ 540.166987] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.166987] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] self.wait() [ 540.166987] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.166987] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] self[:] = self._gt.wait() [ 540.166987] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.166987] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] return self._exit_event.wait() [ 540.166987] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.166987] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] result = hub.switch() [ 540.166987] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.166987] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] return self.greenlet.switch() [ 540.167603] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.167603] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] result = function(*args, **kwargs) [ 540.167603] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.167603] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] return func(*args, **kwargs) [ 540.167603] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.167603] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] raise e [ 540.167603] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.167603] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] nwinfo = self.network_api.allocate_for_instance( [ 540.167603] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.167603] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] created_port_ids = self._update_ports_for_instance( [ 540.167603] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.167603] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] with excutils.save_and_reraise_exception(): [ 540.167603] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.168336] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] self.force_reraise() [ 540.168336] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.168336] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] raise self.value [ 540.168336] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.168336] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] updated_port = self._update_port( [ 540.168336] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.168336] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] _ensure_no_port_binding_failure(port) [ 540.168336] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.168336] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] raise exception.PortBindingFailed(port_id=port['id']) [ 540.168336] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] nova.exception.PortBindingFailed: Binding failed for port 6590d44b-a55c-49b4-8105-3915d4a6d5a2, please check neutron logs for more information. [ 540.168336] env[63372]: ERROR nova.compute.manager [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] [ 540.168708] env[63372]: DEBUG nova.compute.utils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Binding failed for port 6590d44b-a55c-49b4-8105-3915d4a6d5a2, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 540.168708] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 3.636s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.172450] env[63372]: DEBUG nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Build of instance f33eea3e-887c-4164-98e9-e4d502d14af2 was re-scheduled: Binding failed for port 6590d44b-a55c-49b4-8105-3915d4a6d5a2, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 540.172792] env[63372]: DEBUG nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 540.172930] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Acquiring lock "refresh_cache-f33eea3e-887c-4164-98e9-e4d502d14af2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.173074] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Acquired lock "refresh_cache-f33eea3e-887c-4164-98e9-e4d502d14af2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.173233] env[63372]: DEBUG nova.network.neutron [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 540.249317] env[63372]: DEBUG nova.network.neutron [req-ba4fbfb6-c299-45e0-8c40-caf648d15146 req-d6fd5117-fd6b-47c3-a8a5-3b71fc113bb2 service nova] [instance: 32dabedb-8155-4377-871b-408935db97b5] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.396178] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.425987] env[63372]: DEBUG oslo_concurrency.lockutils [req-fe5832f0-73d0-4276-8ca7-c93ea7561df9 req-0a2d85ea-accc-4ddc-8295-6abad0d4752c service nova] Releasing lock "refresh_cache-393f3218-5e22-4e88-8968-9157aa0631da" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.426287] env[63372]: DEBUG nova.compute.manager [req-fe5832f0-73d0-4276-8ca7-c93ea7561df9 req-0a2d85ea-accc-4ddc-8295-6abad0d4752c service nova] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Received event network-vif-deleted-5f466550-60f4-48bb-871d-060148bd4558 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 540.739671] env[63372]: DEBUG nova.network.neutron [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.754536] env[63372]: DEBUG oslo_concurrency.lockutils [req-ba4fbfb6-c299-45e0-8c40-caf648d15146 req-d6fd5117-fd6b-47c3-a8a5-3b71fc113bb2 service nova] Releasing lock "refresh_cache-32dabedb-8155-4377-871b-408935db97b5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.754952] env[63372]: DEBUG nova.compute.manager [req-ba4fbfb6-c299-45e0-8c40-caf648d15146 req-d6fd5117-fd6b-47c3-a8a5-3b71fc113bb2 service nova] [instance: 32dabedb-8155-4377-871b-408935db97b5] Received event network-vif-deleted-8c633cc9-a20a-4af4-ad1f-b367ad5c9fa9 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 540.865144] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8433121-3b3d-418b-9fb3-5e8362980e2a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.885369] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee0632af-5840-4446-9315-ff36945ae3b0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.917733] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c6f709-5792-4cdd-93e2-7a91c04851e9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.927590] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-245b1b83-eaf2-4080-b2bb-040c024e59f2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.936825] env[63372]: DEBUG nova.network.neutron [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.950470] env[63372]: DEBUG nova.compute.provider_tree [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.993498] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Acquiring lock "292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.993639] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Lock "292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.094299] env[63372]: ERROR nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 69372e7a-1cad-4678-8591-897c91ccc7e1, please check neutron logs for more information. [ 541.094299] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 541.094299] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.094299] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 541.094299] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 541.094299] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 541.094299] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 541.094299] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 541.094299] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.094299] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 541.094299] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.094299] env[63372]: ERROR nova.compute.manager raise self.value [ 541.094299] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 541.094299] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 541.094299] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.094299] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 541.094978] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.094978] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 541.094978] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 69372e7a-1cad-4678-8591-897c91ccc7e1, please check neutron logs for more information. [ 541.094978] env[63372]: ERROR nova.compute.manager [ 541.094978] env[63372]: Traceback (most recent call last): [ 541.094978] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 541.094978] env[63372]: listener.cb(fileno) [ 541.094978] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.094978] env[63372]: result = function(*args, **kwargs) [ 541.094978] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 541.094978] env[63372]: return func(*args, **kwargs) [ 541.094978] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.094978] env[63372]: raise e [ 541.094978] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.094978] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 541.094978] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 541.094978] env[63372]: created_port_ids = self._update_ports_for_instance( [ 541.094978] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 541.094978] env[63372]: with excutils.save_and_reraise_exception(): [ 541.094978] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.094978] env[63372]: self.force_reraise() [ 541.094978] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.094978] env[63372]: raise self.value [ 541.094978] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 541.094978] env[63372]: updated_port = self._update_port( [ 541.094978] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.094978] env[63372]: _ensure_no_port_binding_failure(port) [ 541.094978] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.094978] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 541.095894] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 69372e7a-1cad-4678-8591-897c91ccc7e1, please check neutron logs for more information. [ 541.095894] env[63372]: Removing descriptor: 16 [ 541.095894] env[63372]: ERROR nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 69372e7a-1cad-4678-8591-897c91ccc7e1, please check neutron logs for more information. [ 541.095894] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Traceback (most recent call last): [ 541.095894] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 541.095894] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] yield resources [ 541.095894] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 541.095894] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] self.driver.spawn(context, instance, image_meta, [ 541.095894] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 541.095894] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] self._vmops.spawn(context, instance, image_meta, injected_files, [ 541.095894] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 541.095894] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] vm_ref = self.build_virtual_machine(instance, [ 541.096270] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 541.096270] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] vif_infos = vmwarevif.get_vif_info(self._session, [ 541.096270] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 541.096270] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] for vif in network_info: [ 541.096270] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 541.096270] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] return self._sync_wrapper(fn, *args, **kwargs) [ 541.096270] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 541.096270] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] self.wait() [ 541.096270] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 541.096270] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] self[:] = self._gt.wait() [ 541.096270] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 541.096270] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] return self._exit_event.wait() [ 541.096270] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 541.096662] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] result = hub.switch() [ 541.096662] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 541.096662] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] return self.greenlet.switch() [ 541.096662] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.096662] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] result = function(*args, **kwargs) [ 541.096662] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 541.096662] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] return func(*args, **kwargs) [ 541.096662] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.096662] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] raise e [ 541.096662] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.096662] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] nwinfo = self.network_api.allocate_for_instance( [ 541.096662] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 541.096662] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] created_port_ids = self._update_ports_for_instance( [ 541.097020] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 541.097020] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] with excutils.save_and_reraise_exception(): [ 541.097020] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.097020] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] self.force_reraise() [ 541.097020] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.097020] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] raise self.value [ 541.097020] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 541.097020] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] updated_port = self._update_port( [ 541.097020] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.097020] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] _ensure_no_port_binding_failure(port) [ 541.097020] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.097020] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] raise exception.PortBindingFailed(port_id=port['id']) [ 541.097444] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] nova.exception.PortBindingFailed: Binding failed for port 69372e7a-1cad-4678-8591-897c91ccc7e1, please check neutron logs for more information. [ 541.097444] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] [ 541.097444] env[63372]: INFO nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Terminating instance [ 541.104208] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Acquiring lock "refresh_cache-91351ef7-a282-4f57-b96b-ba0af6d0e074" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.104315] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Acquired lock "refresh_cache-91351ef7-a282-4f57-b96b-ba0af6d0e074" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.104481] env[63372]: DEBUG nova.network.neutron [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 541.440124] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Releasing lock "refresh_cache-f33eea3e-887c-4164-98e9-e4d502d14af2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.441030] env[63372]: DEBUG nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 541.441030] env[63372]: DEBUG nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 541.441030] env[63372]: DEBUG nova.network.neutron [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 541.455513] env[63372]: DEBUG nova.scheduler.client.report [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.464149] env[63372]: DEBUG nova.network.neutron [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.496793] env[63372]: DEBUG nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 541.623947] env[63372]: DEBUG nova.network.neutron [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.742978] env[63372]: DEBUG nova.network.neutron [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.869535] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "ee2f5088-616b-4237-93c9-fac5e0d6aba2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.870050] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "ee2f5088-616b-4237-93c9-fac5e0d6aba2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.899885] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "39e22917-3ea8-457c-9d6e-679a848c3fd3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.900266] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "39e22917-3ea8-457c-9d6e-679a848c3fd3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.939807] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "7803b83f-2944-4e3b-87b1-0d3cea3455ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.939976] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "7803b83f-2944-4e3b-87b1-0d3cea3455ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.962636] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.796s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.963394] env[63372]: ERROR nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5f466550-60f4-48bb-871d-060148bd4558, please check neutron logs for more information. [ 541.963394] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Traceback (most recent call last): [ 541.963394] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 541.963394] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] self.driver.spawn(context, instance, image_meta, [ 541.963394] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 541.963394] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] self._vmops.spawn(context, instance, image_meta, injected_files, [ 541.963394] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 541.963394] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] vm_ref = self.build_virtual_machine(instance, [ 541.963394] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 541.963394] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] vif_infos = vmwarevif.get_vif_info(self._session, [ 541.963394] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 541.963693] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] for vif in network_info: [ 541.963693] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 541.963693] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] return self._sync_wrapper(fn, *args, **kwargs) [ 541.963693] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 541.963693] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] self.wait() [ 541.963693] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 541.963693] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] self[:] = self._gt.wait() [ 541.963693] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 541.963693] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] return self._exit_event.wait() [ 541.963693] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 541.963693] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] result = hub.switch() [ 541.963693] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 541.963693] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] return self.greenlet.switch() [ 541.963992] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.963992] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] result = function(*args, **kwargs) [ 541.963992] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 541.963992] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] return func(*args, **kwargs) [ 541.963992] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.963992] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] raise e [ 541.963992] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.963992] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] nwinfo = self.network_api.allocate_for_instance( [ 541.963992] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 541.963992] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] created_port_ids = self._update_ports_for_instance( [ 541.963992] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 541.963992] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] with excutils.save_and_reraise_exception(): [ 541.963992] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.965038] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] self.force_reraise() [ 541.965038] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.965038] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] raise self.value [ 541.965038] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 541.965038] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] updated_port = self._update_port( [ 541.965038] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.965038] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] _ensure_no_port_binding_failure(port) [ 541.965038] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.965038] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] raise exception.PortBindingFailed(port_id=port['id']) [ 541.965038] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] nova.exception.PortBindingFailed: Binding failed for port 5f466550-60f4-48bb-871d-060148bd4558, please check neutron logs for more information. [ 541.965038] env[63372]: ERROR nova.compute.manager [instance: 393f3218-5e22-4e88-8968-9157aa0631da] [ 541.965535] env[63372]: DEBUG nova.compute.utils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Binding failed for port 5f466550-60f4-48bb-871d-060148bd4558, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 541.965997] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 2.732s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.969152] env[63372]: DEBUG nova.network.neutron [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.970735] env[63372]: DEBUG nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Build of instance 393f3218-5e22-4e88-8968-9157aa0631da was re-scheduled: Binding failed for port 5f466550-60f4-48bb-871d-060148bd4558, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 541.971187] env[63372]: DEBUG nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 541.971365] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Acquiring lock "refresh_cache-393f3218-5e22-4e88-8968-9157aa0631da" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.972033] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Acquired lock "refresh_cache-393f3218-5e22-4e88-8968-9157aa0631da" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.972033] env[63372]: DEBUG nova.network.neutron [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 542.035588] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.245704] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Releasing lock "refresh_cache-91351ef7-a282-4f57-b96b-ba0af6d0e074" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.247244] env[63372]: DEBUG nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 542.247340] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 542.247583] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f11be4e5-60bf-4709-aeb0-769165ce13ff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.256803] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d84456-226f-47d1-8ed3-4d1ff192a7a6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.284644] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 91351ef7-a282-4f57-b96b-ba0af6d0e074 could not be found. [ 542.284644] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 542.284644] env[63372]: INFO nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Took 0.04 seconds to destroy the instance on the hypervisor. [ 542.284850] env[63372]: DEBUG oslo.service.loopingcall [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.284988] env[63372]: DEBUG nova.compute.manager [-] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 542.285156] env[63372]: DEBUG nova.network.neutron [-] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 542.305900] env[63372]: DEBUG nova.network.neutron [-] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 542.372836] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 542.403769] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 542.444776] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 542.478021] env[63372]: INFO nova.compute.manager [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] [instance: f33eea3e-887c-4164-98e9-e4d502d14af2] Took 1.03 seconds to deallocate network for instance. [ 542.501799] env[63372]: DEBUG nova.network.neutron [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 542.571641] env[63372]: DEBUG nova.network.neutron [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.694828] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c000845-2b5c-4f92-bac8-13ec474f4565 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.702959] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7edbcf8f-c5f6-46fa-9b29-48393d0113a8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.742965] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0bef2f-7e68-46be-b898-0bd7494063e6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.745917] env[63372]: DEBUG nova.compute.manager [req-d1be77f0-7033-4cb5-9d77-954c0f542b79 req-3d45c3ea-cb74-45d0-8c1b-0d6a2f3be736 service nova] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Received event network-changed-c92db39d-fd16-4e13-a151-e5781476ef44 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.746123] env[63372]: DEBUG nova.compute.manager [req-d1be77f0-7033-4cb5-9d77-954c0f542b79 req-3d45c3ea-cb74-45d0-8c1b-0d6a2f3be736 service nova] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Refreshing instance network info cache due to event network-changed-c92db39d-fd16-4e13-a151-e5781476ef44. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 542.746329] env[63372]: DEBUG oslo_concurrency.lockutils [req-d1be77f0-7033-4cb5-9d77-954c0f542b79 req-3d45c3ea-cb74-45d0-8c1b-0d6a2f3be736 service nova] Acquiring lock "refresh_cache-21c4415c-a016-47eb-9ec4-280c6c18b3fb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.746523] env[63372]: DEBUG oslo_concurrency.lockutils [req-d1be77f0-7033-4cb5-9d77-954c0f542b79 req-3d45c3ea-cb74-45d0-8c1b-0d6a2f3be736 service nova] Acquired lock "refresh_cache-21c4415c-a016-47eb-9ec4-280c6c18b3fb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.746682] env[63372]: DEBUG nova.network.neutron [req-d1be77f0-7033-4cb5-9d77-954c0f542b79 req-3d45c3ea-cb74-45d0-8c1b-0d6a2f3be736 service nova] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Refreshing network info cache for port c92db39d-fd16-4e13-a151-e5781476ef44 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 542.751827] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Acquiring lock "9ec742c1-894f-4baa-a451-01ebb4d2816a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.752078] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Lock "9ec742c1-894f-4baa-a451-01ebb4d2816a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.758294] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b276d2e-f8df-4ad1-9cc1-ce29ee68edd8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.774845] env[63372]: DEBUG nova.compute.provider_tree [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.792464] env[63372]: DEBUG nova.compute.manager [req-aa660392-d3e2-4912-87d2-09beb09ebaa1 req-dc927672-4973-4726-a267-5d7fc876fc98 service nova] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Received event network-changed-69372e7a-1cad-4678-8591-897c91ccc7e1 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.792689] env[63372]: DEBUG nova.compute.manager [req-aa660392-d3e2-4912-87d2-09beb09ebaa1 req-dc927672-4973-4726-a267-5d7fc876fc98 service nova] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Refreshing instance network info cache due to event network-changed-69372e7a-1cad-4678-8591-897c91ccc7e1. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 542.792883] env[63372]: DEBUG oslo_concurrency.lockutils [req-aa660392-d3e2-4912-87d2-09beb09ebaa1 req-dc927672-4973-4726-a267-5d7fc876fc98 service nova] Acquiring lock "refresh_cache-91351ef7-a282-4f57-b96b-ba0af6d0e074" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.793339] env[63372]: DEBUG oslo_concurrency.lockutils [req-aa660392-d3e2-4912-87d2-09beb09ebaa1 req-dc927672-4973-4726-a267-5d7fc876fc98 service nova] Acquired lock "refresh_cache-91351ef7-a282-4f57-b96b-ba0af6d0e074" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.793567] env[63372]: DEBUG nova.network.neutron [req-aa660392-d3e2-4912-87d2-09beb09ebaa1 req-dc927672-4973-4726-a267-5d7fc876fc98 service nova] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Refreshing network info cache for port 69372e7a-1cad-4678-8591-897c91ccc7e1 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 542.808277] env[63372]: DEBUG nova.network.neutron [-] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.903717] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.923399] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.966254] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.074852] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Releasing lock "refresh_cache-393f3218-5e22-4e88-8968-9157aa0631da" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.074852] env[63372]: DEBUG nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 543.075168] env[63372]: DEBUG nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 543.075168] env[63372]: DEBUG nova.network.neutron [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 543.097796] env[63372]: DEBUG nova.network.neutron [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.278302] env[63372]: DEBUG nova.scheduler.client.report [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.287512] env[63372]: DEBUG nova.network.neutron [req-d1be77f0-7033-4cb5-9d77-954c0f542b79 req-3d45c3ea-cb74-45d0-8c1b-0d6a2f3be736 service nova] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.313840] env[63372]: INFO nova.compute.manager [-] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Took 1.03 seconds to deallocate network for instance. [ 543.314973] env[63372]: DEBUG nova.compute.claims [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 543.315202] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.334356] env[63372]: DEBUG nova.network.neutron [req-aa660392-d3e2-4912-87d2-09beb09ebaa1 req-dc927672-4973-4726-a267-5d7fc876fc98 service nova] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.496662] env[63372]: DEBUG nova.network.neutron [req-d1be77f0-7033-4cb5-9d77-954c0f542b79 req-3d45c3ea-cb74-45d0-8c1b-0d6a2f3be736 service nova] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.518858] env[63372]: INFO nova.scheduler.client.report [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Deleted allocations for instance f33eea3e-887c-4164-98e9-e4d502d14af2 [ 543.548962] env[63372]: DEBUG nova.network.neutron [req-aa660392-d3e2-4912-87d2-09beb09ebaa1 req-dc927672-4973-4726-a267-5d7fc876fc98 service nova] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.602432] env[63372]: DEBUG nova.network.neutron [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.786979] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.820s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.786979] env[63372]: ERROR nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8c633cc9-a20a-4af4-ad1f-b367ad5c9fa9, please check neutron logs for more information. [ 543.786979] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] Traceback (most recent call last): [ 543.786979] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 543.786979] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] self.driver.spawn(context, instance, image_meta, [ 543.786979] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 543.786979] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.786979] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.786979] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] vm_ref = self.build_virtual_machine(instance, [ 543.787402] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.787402] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.787402] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.787402] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] for vif in network_info: [ 543.787402] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 543.787402] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] return self._sync_wrapper(fn, *args, **kwargs) [ 543.787402] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 543.787402] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] self.wait() [ 543.787402] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 543.787402] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] self[:] = self._gt.wait() [ 543.787402] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.787402] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] return self._exit_event.wait() [ 543.787402] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.787720] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] result = hub.switch() [ 543.787720] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.787720] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] return self.greenlet.switch() [ 543.787720] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.787720] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] result = function(*args, **kwargs) [ 543.787720] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 543.787720] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] return func(*args, **kwargs) [ 543.787720] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.787720] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] raise e [ 543.787720] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.787720] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] nwinfo = self.network_api.allocate_for_instance( [ 543.787720] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 543.787720] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] created_port_ids = self._update_ports_for_instance( [ 543.788068] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 543.788068] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] with excutils.save_and_reraise_exception(): [ 543.788068] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.788068] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] self.force_reraise() [ 543.788068] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.788068] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] raise self.value [ 543.788068] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 543.788068] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] updated_port = self._update_port( [ 543.788068] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.788068] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] _ensure_no_port_binding_failure(port) [ 543.788068] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.788068] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] raise exception.PortBindingFailed(port_id=port['id']) [ 543.788362] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] nova.exception.PortBindingFailed: Binding failed for port 8c633cc9-a20a-4af4-ad1f-b367ad5c9fa9, please check neutron logs for more information. [ 543.788362] env[63372]: ERROR nova.compute.manager [instance: 32dabedb-8155-4377-871b-408935db97b5] [ 543.788362] env[63372]: DEBUG nova.compute.utils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Binding failed for port 8c633cc9-a20a-4af4-ad1f-b367ad5c9fa9, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 543.789616] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 3.790s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.792925] env[63372]: DEBUG nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Build of instance 32dabedb-8155-4377-871b-408935db97b5 was re-scheduled: Binding failed for port 8c633cc9-a20a-4af4-ad1f-b367ad5c9fa9, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 543.793849] env[63372]: DEBUG nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 543.793849] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Acquiring lock "refresh_cache-32dabedb-8155-4377-871b-408935db97b5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.793849] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Acquired lock "refresh_cache-32dabedb-8155-4377-871b-408935db97b5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.794125] env[63372]: DEBUG nova.network.neutron [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 544.002377] env[63372]: DEBUG oslo_concurrency.lockutils [req-d1be77f0-7033-4cb5-9d77-954c0f542b79 req-3d45c3ea-cb74-45d0-8c1b-0d6a2f3be736 service nova] Releasing lock "refresh_cache-21c4415c-a016-47eb-9ec4-280c6c18b3fb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.002377] env[63372]: DEBUG nova.compute.manager [req-d1be77f0-7033-4cb5-9d77-954c0f542b79 req-3d45c3ea-cb74-45d0-8c1b-0d6a2f3be736 service nova] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Received event network-vif-deleted-c92db39d-fd16-4e13-a151-e5781476ef44 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.036054] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8bcd4536-645e-4b9f-913d-b44998c2ed36 tempest-TenantUsagesTestJSON-1875412056 tempest-TenantUsagesTestJSON-1875412056-project-member] Lock "f33eea3e-887c-4164-98e9-e4d502d14af2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.688s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.057332] env[63372]: DEBUG oslo_concurrency.lockutils [req-aa660392-d3e2-4912-87d2-09beb09ebaa1 req-dc927672-4973-4726-a267-5d7fc876fc98 service nova] Releasing lock "refresh_cache-91351ef7-a282-4f57-b96b-ba0af6d0e074" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.058360] env[63372]: DEBUG nova.compute.manager [req-aa660392-d3e2-4912-87d2-09beb09ebaa1 req-dc927672-4973-4726-a267-5d7fc876fc98 service nova] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Received event network-vif-deleted-69372e7a-1cad-4678-8591-897c91ccc7e1 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.108088] env[63372]: INFO nova.compute.manager [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] [instance: 393f3218-5e22-4e88-8968-9157aa0631da] Took 1.03 seconds to deallocate network for instance. [ 544.345309] env[63372]: DEBUG nova.network.neutron [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.470861] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d578d898-55e3-4206-8d13-b08c0cca0c6d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.478433] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a5550b-aeb6-41ae-8f7e-17761ff6c8cd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.524694] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1189ac-8c4e-4dc2-90bc-ab0ff9610274 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.533663] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeedf833-2b1d-4b4d-be02-8417f3dd1dcc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.538394] env[63372]: DEBUG nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 544.552417] env[63372]: DEBUG nova.compute.provider_tree [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.822644] env[63372]: DEBUG nova.network.neutron [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.881298] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Acquiring lock "dc9c4a56-e0af-473b-b47b-2061fe89c6d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.881541] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Lock "dc9c4a56-e0af-473b-b47b-2061fe89c6d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.056746] env[63372]: DEBUG nova.scheduler.client.report [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 545.071232] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.155134] env[63372]: INFO nova.scheduler.client.report [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Deleted allocations for instance 393f3218-5e22-4e88-8968-9157aa0631da [ 545.327700] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Releasing lock "refresh_cache-32dabedb-8155-4377-871b-408935db97b5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.327700] env[63372]: DEBUG nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 545.327700] env[63372]: DEBUG nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.327700] env[63372]: DEBUG nova.network.neutron [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 545.373019] env[63372]: DEBUG nova.network.neutron [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.563910] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.774s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.566033] env[63372]: ERROR nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c92db39d-fd16-4e13-a151-e5781476ef44, please check neutron logs for more information. [ 545.566033] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Traceback (most recent call last): [ 545.566033] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 545.566033] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] self.driver.spawn(context, instance, image_meta, [ 545.566033] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 545.566033] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.566033] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.566033] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] vm_ref = self.build_virtual_machine(instance, [ 545.566033] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.566033] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.566033] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.566432] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] for vif in network_info: [ 545.566432] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.566432] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] return self._sync_wrapper(fn, *args, **kwargs) [ 545.566432] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.566432] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] self.wait() [ 545.566432] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.566432] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] self[:] = self._gt.wait() [ 545.566432] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.566432] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] return self._exit_event.wait() [ 545.566432] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.566432] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] result = hub.switch() [ 545.566432] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.566432] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] return self.greenlet.switch() [ 545.566811] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.566811] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] result = function(*args, **kwargs) [ 545.566811] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.566811] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] return func(*args, **kwargs) [ 545.566811] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.566811] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] raise e [ 545.566811] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.566811] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] nwinfo = self.network_api.allocate_for_instance( [ 545.566811] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.566811] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] created_port_ids = self._update_ports_for_instance( [ 545.566811] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.566811] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] with excutils.save_and_reraise_exception(): [ 545.566811] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.567185] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] self.force_reraise() [ 545.567185] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.567185] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] raise self.value [ 545.567185] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.567185] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] updated_port = self._update_port( [ 545.567185] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.567185] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] _ensure_no_port_binding_failure(port) [ 545.567185] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.567185] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] raise exception.PortBindingFailed(port_id=port['id']) [ 545.567185] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] nova.exception.PortBindingFailed: Binding failed for port c92db39d-fd16-4e13-a151-e5781476ef44, please check neutron logs for more information. [ 545.567185] env[63372]: ERROR nova.compute.manager [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] [ 545.569091] env[63372]: DEBUG nova.compute.utils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Binding failed for port c92db39d-fd16-4e13-a151-e5781476ef44, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 545.570688] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.175s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.575762] env[63372]: INFO nova.compute.claims [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 545.583666] env[63372]: DEBUG nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Build of instance 21c4415c-a016-47eb-9ec4-280c6c18b3fb was re-scheduled: Binding failed for port c92db39d-fd16-4e13-a151-e5781476ef44, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 545.584273] env[63372]: DEBUG nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 545.584502] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "refresh_cache-21c4415c-a016-47eb-9ec4-280c6c18b3fb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.584637] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired lock "refresh_cache-21c4415c-a016-47eb-9ec4-280c6c18b3fb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.584792] env[63372]: DEBUG nova.network.neutron [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 545.664383] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5784113-9639-4989-9d56-317a782b78eb tempest-ServerDiagnosticsNegativeTest-477537969 tempest-ServerDiagnosticsNegativeTest-477537969-project-member] Lock "393f3218-5e22-4e88-8968-9157aa0631da" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.550s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.875311] env[63372]: DEBUG nova.network.neutron [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.134224] env[63372]: DEBUG nova.network.neutron [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.169319] env[63372]: DEBUG nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 546.286385] env[63372]: DEBUG nova.network.neutron [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.377667] env[63372]: INFO nova.compute.manager [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] [instance: 32dabedb-8155-4377-871b-408935db97b5] Took 1.05 seconds to deallocate network for instance. [ 546.493172] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Acquiring lock "392f418f-c8c7-41e3-9bb6-f68e1cab75f3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.493172] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Lock "392f418f-c8c7-41e3-9bb6-f68e1cab75f3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.698896] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.788696] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Releasing lock "refresh_cache-21c4415c-a016-47eb-9ec4-280c6c18b3fb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.788980] env[63372]: DEBUG nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 546.789127] env[63372]: DEBUG nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 546.789284] env[63372]: DEBUG nova.network.neutron [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 546.794421] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ad373a-a270-4be2-90b4-e7b2ab149683 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.803672] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a855fd5-89b0-4c28-b104-bb06231f5375 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.838048] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e90d5aa-8373-49f3-9e20-2b7901b98351 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.845020] env[63372]: DEBUG nova.network.neutron [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.847842] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf20822-5a11-4c30-b013-ab9019c1ad9c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.855261] env[63372]: DEBUG nova.network.neutron [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.867808] env[63372]: DEBUG nova.compute.provider_tree [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.359440] env[63372]: INFO nova.compute.manager [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 21c4415c-a016-47eb-9ec4-280c6c18b3fb] Took 0.57 seconds to deallocate network for instance. [ 547.371227] env[63372]: DEBUG nova.scheduler.client.report [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 547.419140] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "9d52ebd0-f38e-4825-a974-f9a0c0643116" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.420044] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "9d52ebd0-f38e-4825-a974-f9a0c0643116" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.431354] env[63372]: INFO nova.scheduler.client.report [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Deleted allocations for instance 32dabedb-8155-4377-871b-408935db97b5 [ 547.879076] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.881270] env[63372]: DEBUG nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 547.886567] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.851s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.888161] env[63372]: INFO nova.compute.claims [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 547.945646] env[63372]: DEBUG oslo_concurrency.lockutils [None req-004fabc5-e8ac-422a-8821-ae4c561ec960 tempest-FloatingIPsAssociationTestJSON-2132179673 tempest-FloatingIPsAssociationTestJSON-2132179673-project-member] Lock "32dabedb-8155-4377-871b-408935db97b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.387s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.011039] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Acquiring lock "45d6d393-ca23-4e22-bcbf-899ae5037c4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.011039] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Lock "45d6d393-ca23-4e22-bcbf-899ae5037c4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.396038] env[63372]: DEBUG nova.compute.utils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 548.403923] env[63372]: DEBUG nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 548.403923] env[63372]: DEBUG nova.network.neutron [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 548.411080] env[63372]: INFO nova.scheduler.client.report [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Deleted allocations for instance 21c4415c-a016-47eb-9ec4-280c6c18b3fb [ 548.451260] env[63372]: DEBUG nova.policy [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'feb3c6c6987e4a70a7c9588793ff52eb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ed6d958c3204cd8b88fdb363d6752ee', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 548.454279] env[63372]: DEBUG nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.904339] env[63372]: DEBUG nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 548.924735] env[63372]: DEBUG oslo_concurrency.lockutils [None req-790cfb9e-c511-43f2-860e-910ab17e5244 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "21c4415c-a016-47eb-9ec4-280c6c18b3fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.263s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.984461] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.157622] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad2a860-2c29-406c-86d2-744b7126db96 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.166495] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f328103-45b0-4f91-ac26-2ab7941b8dd1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.199100] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5061a2-f21b-4908-90c7-23bdac12ad64 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.207287] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4a09af-527a-4def-b9bf-63ed2f90dbf7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.224531] env[63372]: DEBUG nova.compute.provider_tree [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.409029] env[63372]: DEBUG nova.network.neutron [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Successfully created port: d0f239ea-5afd-491e-a640-c038817aee6a {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 549.430121] env[63372]: DEBUG nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 549.727946] env[63372]: DEBUG nova.scheduler.client.report [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 549.923963] env[63372]: DEBUG nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 549.964230] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.967144] env[63372]: DEBUG nova.virt.hardware [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 549.968126] env[63372]: DEBUG nova.virt.hardware [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 549.968126] env[63372]: DEBUG nova.virt.hardware [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 549.968126] env[63372]: DEBUG nova.virt.hardware [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 549.968126] env[63372]: DEBUG nova.virt.hardware [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 549.968126] env[63372]: DEBUG nova.virt.hardware [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 549.968372] env[63372]: DEBUG nova.virt.hardware [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 549.968478] env[63372]: DEBUG nova.virt.hardware [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 549.968790] env[63372]: DEBUG nova.virt.hardware [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 549.968790] env[63372]: DEBUG nova.virt.hardware [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 549.968963] env[63372]: DEBUG nova.virt.hardware [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 549.969852] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7287a188-e4a8-4aa2-8315-b51d1e23d5e5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.984348] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b1c112-94c8-4829-875c-8335f0980ba6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.237776] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.351s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.238204] env[63372]: DEBUG nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 550.244506] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.338s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.244506] env[63372]: INFO nova.compute.claims [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 550.750506] env[63372]: DEBUG nova.compute.utils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 550.752955] env[63372]: DEBUG nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 550.753127] env[63372]: DEBUG nova.network.neutron [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 550.981663] env[63372]: DEBUG nova.policy [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ca2520544ca45b3b962a44570cfd25b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32fe239916e34149be0c064d73ea4de5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 551.075463] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Acquiring lock "fe96690d-4ca0-413b-9e04-09e7f73a0bb7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.075729] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Lock "fe96690d-4ca0-413b-9e04-09e7f73a0bb7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.255569] env[63372]: DEBUG nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 551.521769] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3bd9cc-7694-4653-9c74-ef612e2a8b1e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.533594] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7acacd6b-028d-4929-8443-f9c05cb25593 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.576523] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10571b25-eb97-4abe-93db-a29fe55f97ec {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.584026] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9a9dea-ae27-46cb-a88b-a1be653f990a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.602979] env[63372]: DEBUG nova.compute.provider_tree [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.770352] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "d0310e79-e07a-4581-8cd8-34437cd9ea7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.771210] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "d0310e79-e07a-4581-8cd8-34437cd9ea7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.109153] env[63372]: DEBUG nova.scheduler.client.report [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 552.114993] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Acquiring lock "6a6268e4-0539-45cf-9335-cde31473d1f9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.115455] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Lock "6a6268e4-0539-45cf-9335-cde31473d1f9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.270124] env[63372]: DEBUG nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 552.309563] env[63372]: DEBUG nova.virt.hardware [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 552.309817] env[63372]: DEBUG nova.virt.hardware [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 552.310483] env[63372]: DEBUG nova.virt.hardware [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 552.310756] env[63372]: DEBUG nova.virt.hardware [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 552.310917] env[63372]: DEBUG nova.virt.hardware [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 552.311079] env[63372]: DEBUG nova.virt.hardware [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 552.311300] env[63372]: DEBUG nova.virt.hardware [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 552.311457] env[63372]: DEBUG nova.virt.hardware [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 552.311618] env[63372]: DEBUG nova.virt.hardware [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 552.312365] env[63372]: DEBUG nova.virt.hardware [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 552.312406] env[63372]: DEBUG nova.virt.hardware [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 552.316701] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6afb01e-f034-49e1-97be-e6a5f7210f02 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.327052] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfe40c6-6e5c-4a80-b19e-9c05b790c409 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.450484] env[63372]: DEBUG nova.network.neutron [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Successfully created port: 2bdb48e9-8cdf-4e6d-b8ca-12d83d6a2b26 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 552.616725] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.618262] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 552.621513] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.698s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.623401] env[63372]: INFO nova.compute.claims [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 552.979862] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Acquiring lock "96c3ce22-9424-4f7f-8810-ee7cfaed0193" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.980187] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Lock "96c3ce22-9424-4f7f-8810-ee7cfaed0193" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.130099] env[63372]: DEBUG nova.compute.utils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 553.133393] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 553.133587] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 553.332420] env[63372]: DEBUG nova.policy [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5a6bf5260f334ed687da2c419433af73', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78fd6c4fc305495abb4a64e0c26b75db', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 553.641820] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 553.968632] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa9e95d-6e49-41c4-8930-da1d6bab6fe5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.980552] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63617ad2-6cb6-452d-86d6-e34ad156f437 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.013405] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ddeea55-28ca-4559-9473-61aa7c2b7303 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.021624] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56077507-bb79-459e-9574-2511b7f5eb1b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.039260] env[63372]: DEBUG nova.compute.provider_tree [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 554.547186] env[63372]: DEBUG nova.scheduler.client.report [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 554.590265] env[63372]: ERROR nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d0f239ea-5afd-491e-a640-c038817aee6a, please check neutron logs for more information. [ 554.590265] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 554.590265] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.590265] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 554.590265] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.590265] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 554.590265] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.590265] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 554.590265] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.590265] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 554.590265] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.590265] env[63372]: ERROR nova.compute.manager raise self.value [ 554.590265] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.590265] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 554.590265] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.590265] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 554.590952] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.590952] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 554.590952] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d0f239ea-5afd-491e-a640-c038817aee6a, please check neutron logs for more information. [ 554.590952] env[63372]: ERROR nova.compute.manager [ 554.590952] env[63372]: Traceback (most recent call last): [ 554.590952] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 554.590952] env[63372]: listener.cb(fileno) [ 554.590952] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.590952] env[63372]: result = function(*args, **kwargs) [ 554.590952] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.590952] env[63372]: return func(*args, **kwargs) [ 554.590952] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.590952] env[63372]: raise e [ 554.590952] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.590952] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 554.590952] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.590952] env[63372]: created_port_ids = self._update_ports_for_instance( [ 554.590952] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.590952] env[63372]: with excutils.save_and_reraise_exception(): [ 554.590952] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.590952] env[63372]: self.force_reraise() [ 554.590952] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.590952] env[63372]: raise self.value [ 554.590952] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.590952] env[63372]: updated_port = self._update_port( [ 554.590952] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.590952] env[63372]: _ensure_no_port_binding_failure(port) [ 554.590952] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.590952] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 554.591864] env[63372]: nova.exception.PortBindingFailed: Binding failed for port d0f239ea-5afd-491e-a640-c038817aee6a, please check neutron logs for more information. [ 554.591864] env[63372]: Removing descriptor: 18 [ 554.591864] env[63372]: ERROR nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d0f239ea-5afd-491e-a640-c038817aee6a, please check neutron logs for more information. [ 554.591864] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Traceback (most recent call last): [ 554.591864] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 554.591864] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] yield resources [ 554.591864] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 554.591864] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] self.driver.spawn(context, instance, image_meta, [ 554.591864] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 554.591864] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 554.591864] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 554.591864] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] vm_ref = self.build_virtual_machine(instance, [ 554.592296] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 554.592296] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] vif_infos = vmwarevif.get_vif_info(self._session, [ 554.592296] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 554.592296] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] for vif in network_info: [ 554.592296] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 554.592296] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] return self._sync_wrapper(fn, *args, **kwargs) [ 554.592296] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 554.592296] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] self.wait() [ 554.592296] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 554.592296] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] self[:] = self._gt.wait() [ 554.592296] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 554.592296] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] return self._exit_event.wait() [ 554.592296] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 554.592664] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] result = hub.switch() [ 554.592664] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 554.592664] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] return self.greenlet.switch() [ 554.592664] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.592664] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] result = function(*args, **kwargs) [ 554.592664] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.592664] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] return func(*args, **kwargs) [ 554.592664] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.592664] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] raise e [ 554.592664] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.592664] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] nwinfo = self.network_api.allocate_for_instance( [ 554.592664] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.592664] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] created_port_ids = self._update_ports_for_instance( [ 554.593039] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.593039] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] with excutils.save_and_reraise_exception(): [ 554.593039] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.593039] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] self.force_reraise() [ 554.593039] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.593039] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] raise self.value [ 554.593039] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.593039] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] updated_port = self._update_port( [ 554.593039] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.593039] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] _ensure_no_port_binding_failure(port) [ 554.593039] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.593039] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] raise exception.PortBindingFailed(port_id=port['id']) [ 554.593522] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] nova.exception.PortBindingFailed: Binding failed for port d0f239ea-5afd-491e-a640-c038817aee6a, please check neutron logs for more information. [ 554.593522] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] [ 554.593522] env[63372]: INFO nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Terminating instance [ 554.593522] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Acquiring lock "refresh_cache-d290c8b5-422b-43c6-825e-52ea98516e6c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.593522] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Acquired lock "refresh_cache-d290c8b5-422b-43c6-825e-52ea98516e6c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.593522] env[63372]: DEBUG nova.network.neutron [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 554.663628] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 554.699409] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 554.700349] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 554.700349] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 554.700349] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 554.700775] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 554.700775] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 554.700960] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 554.701150] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 554.701358] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 554.701575] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 554.701781] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 554.702739] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24643bad-5a50-405c-8df8-08b64194f757 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.712482] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2aa1c8-64c6-427a-b20a-d36fc2d54617 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.808490] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Acquiring lock "c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.809100] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Lock "c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.012756] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Successfully created port: 1e3d9db7-b8f0-4221-9a0f-8538208cb01b {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 555.053019] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.431s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.053511] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 555.056244] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.090s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.059109] env[63372]: INFO nova.compute.claims [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 555.164264] env[63372]: DEBUG nova.network.neutron [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.416734] env[63372]: DEBUG nova.compute.manager [req-0a737f48-aa5f-46be-af94-1ca598db74b7 req-5766df77-b711-4166-b3ed-83871b76292f service nova] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Received event network-changed-d0f239ea-5afd-491e-a640-c038817aee6a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 555.416955] env[63372]: DEBUG nova.compute.manager [req-0a737f48-aa5f-46be-af94-1ca598db74b7 req-5766df77-b711-4166-b3ed-83871b76292f service nova] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Refreshing instance network info cache due to event network-changed-d0f239ea-5afd-491e-a640-c038817aee6a. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 555.417127] env[63372]: DEBUG oslo_concurrency.lockutils [req-0a737f48-aa5f-46be-af94-1ca598db74b7 req-5766df77-b711-4166-b3ed-83871b76292f service nova] Acquiring lock "refresh_cache-d290c8b5-422b-43c6-825e-52ea98516e6c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.519168] env[63372]: DEBUG nova.network.neutron [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.564251] env[63372]: DEBUG nova.compute.utils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 555.569562] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 555.569730] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 555.752218] env[63372]: DEBUG nova.policy [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5a6bf5260f334ed687da2c419433af73', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78fd6c4fc305495abb4a64e0c26b75db', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 555.851533] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Acquiring lock "8986e782-4d63-4491-8100-4341a3149812" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.851633] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Lock "8986e782-4d63-4491-8100-4341a3149812" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.026164] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Releasing lock "refresh_cache-d290c8b5-422b-43c6-825e-52ea98516e6c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.026164] env[63372]: DEBUG nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 556.026699] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 556.026699] env[63372]: DEBUG oslo_concurrency.lockutils [req-0a737f48-aa5f-46be-af94-1ca598db74b7 req-5766df77-b711-4166-b3ed-83871b76292f service nova] Acquired lock "refresh_cache-d290c8b5-422b-43c6-825e-52ea98516e6c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.026837] env[63372]: DEBUG nova.network.neutron [req-0a737f48-aa5f-46be-af94-1ca598db74b7 req-5766df77-b711-4166-b3ed-83871b76292f service nova] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Refreshing network info cache for port d0f239ea-5afd-491e-a640-c038817aee6a {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 556.030396] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e63dfe1d-4d34-4962-b21d-ae3f446c1c6a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.042668] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e56a9f3-d291-4d71-b9d4-b3cd1701f867 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.066365] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d290c8b5-422b-43c6-825e-52ea98516e6c could not be found. [ 556.069791] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 556.069791] env[63372]: INFO nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 556.069791] env[63372]: DEBUG oslo.service.loopingcall [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 556.069791] env[63372]: DEBUG nova.compute.manager [-] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 556.069791] env[63372]: DEBUG nova.network.neutron [-] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 556.070718] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 556.126050] env[63372]: DEBUG nova.network.neutron [-] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.377342] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5324a16b-c158-46ae-a807-933b75e5b369 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.389117] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fdd24a-95c1-412b-9b97-5bbf263c5266 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.421671] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88381323-eeec-4c0c-a748-438f2a4cb72a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.429304] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1813bf-385d-420c-b4ef-1c17836fde69 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.442904] env[63372]: DEBUG nova.compute.provider_tree [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.595465] env[63372]: DEBUG nova.network.neutron [req-0a737f48-aa5f-46be-af94-1ca598db74b7 req-5766df77-b711-4166-b3ed-83871b76292f service nova] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.632103] env[63372]: DEBUG nova.network.neutron [-] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.691196] env[63372]: DEBUG oslo_concurrency.lockutils [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquiring lock "715408b8-0e76-4b61-a342-b168377cb288" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.691603] env[63372]: DEBUG oslo_concurrency.lockutils [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Lock "715408b8-0e76-4b61-a342-b168377cb288" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.933854] env[63372]: DEBUG nova.network.neutron [req-0a737f48-aa5f-46be-af94-1ca598db74b7 req-5766df77-b711-4166-b3ed-83871b76292f service nova] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.946160] env[63372]: DEBUG nova.scheduler.client.report [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 557.087459] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 557.116046] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 557.117994] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 557.117994] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 557.122030] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 557.122030] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 557.122030] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 557.122030] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 557.122030] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 557.122240] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 557.122240] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 557.122240] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 557.122240] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584db44f-2457-4a87-a52d-3355a868544f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.130485] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa10f98-bacc-4fa6-88fb-1f9215ad949c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.135529] env[63372]: INFO nova.compute.manager [-] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Took 1.07 seconds to deallocate network for instance. [ 557.138672] env[63372]: DEBUG nova.compute.claims [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 557.139043] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.439527] env[63372]: DEBUG oslo_concurrency.lockutils [req-0a737f48-aa5f-46be-af94-1ca598db74b7 req-5766df77-b711-4166-b3ed-83871b76292f service nova] Releasing lock "refresh_cache-d290c8b5-422b-43c6-825e-52ea98516e6c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.452183] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.452183] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 557.454351] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.139s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.601152] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Successfully created port: 66605f0f-5cf3-4815-9309-0d3bcd4144d8 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 557.923924] env[63372]: ERROR nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2bdb48e9-8cdf-4e6d-b8ca-12d83d6a2b26, please check neutron logs for more information. [ 557.923924] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 557.923924] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.923924] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 557.923924] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.923924] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 557.923924] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.923924] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 557.923924] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.923924] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 557.923924] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.923924] env[63372]: ERROR nova.compute.manager raise self.value [ 557.923924] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.923924] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 557.923924] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.923924] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 557.924379] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.924379] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 557.924379] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2bdb48e9-8cdf-4e6d-b8ca-12d83d6a2b26, please check neutron logs for more information. [ 557.924379] env[63372]: ERROR nova.compute.manager [ 557.924379] env[63372]: Traceback (most recent call last): [ 557.924379] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 557.924379] env[63372]: listener.cb(fileno) [ 557.924379] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.924379] env[63372]: result = function(*args, **kwargs) [ 557.924379] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.924379] env[63372]: return func(*args, **kwargs) [ 557.924379] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.924379] env[63372]: raise e [ 557.924379] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.924379] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 557.924379] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.924379] env[63372]: created_port_ids = self._update_ports_for_instance( [ 557.924379] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.924379] env[63372]: with excutils.save_and_reraise_exception(): [ 557.924379] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.924379] env[63372]: self.force_reraise() [ 557.924379] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.924379] env[63372]: raise self.value [ 557.924379] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.924379] env[63372]: updated_port = self._update_port( [ 557.924379] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.924379] env[63372]: _ensure_no_port_binding_failure(port) [ 557.924379] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.924379] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 557.925128] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 2bdb48e9-8cdf-4e6d-b8ca-12d83d6a2b26, please check neutron logs for more information. [ 557.925128] env[63372]: Removing descriptor: 16 [ 557.925128] env[63372]: ERROR nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2bdb48e9-8cdf-4e6d-b8ca-12d83d6a2b26, please check neutron logs for more information. [ 557.925128] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Traceback (most recent call last): [ 557.925128] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 557.925128] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] yield resources [ 557.925128] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.925128] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] self.driver.spawn(context, instance, image_meta, [ 557.925128] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 557.925128] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.925128] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.925128] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] vm_ref = self.build_virtual_machine(instance, [ 557.925595] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.925595] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.925595] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.925595] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] for vif in network_info: [ 557.925595] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.925595] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] return self._sync_wrapper(fn, *args, **kwargs) [ 557.925595] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.925595] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] self.wait() [ 557.925595] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.925595] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] self[:] = self._gt.wait() [ 557.925595] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.925595] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] return self._exit_event.wait() [ 557.925595] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.926732] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] result = hub.switch() [ 557.926732] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.926732] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] return self.greenlet.switch() [ 557.926732] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.926732] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] result = function(*args, **kwargs) [ 557.926732] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.926732] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] return func(*args, **kwargs) [ 557.926732] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.926732] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] raise e [ 557.926732] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.926732] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] nwinfo = self.network_api.allocate_for_instance( [ 557.926732] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.926732] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] created_port_ids = self._update_ports_for_instance( [ 557.928112] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.928112] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] with excutils.save_and_reraise_exception(): [ 557.928112] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.928112] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] self.force_reraise() [ 557.928112] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.928112] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] raise self.value [ 557.928112] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.928112] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] updated_port = self._update_port( [ 557.928112] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.928112] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] _ensure_no_port_binding_failure(port) [ 557.928112] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.928112] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] raise exception.PortBindingFailed(port_id=port['id']) [ 557.928892] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] nova.exception.PortBindingFailed: Binding failed for port 2bdb48e9-8cdf-4e6d-b8ca-12d83d6a2b26, please check neutron logs for more information. [ 557.928892] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] [ 557.928892] env[63372]: INFO nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Terminating instance [ 557.929639] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Acquiring lock "refresh_cache-292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.929813] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Acquired lock "refresh_cache-292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.929973] env[63372]: DEBUG nova.network.neutron [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 557.957898] env[63372]: DEBUG nova.compute.utils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 557.958545] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 557.961418] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 558.162244] env[63372]: DEBUG nova.policy [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5a6bf5260f334ed687da2c419433af73', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78fd6c4fc305495abb4a64e0c26b75db', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 558.292983] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e26b59-e773-4e6b-a94a-9ef0c59038ad {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.304020] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ff1203-d519-4a13-a437-12d90cf592b5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.335618] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ebdb46a-9bc9-48e3-9ee6-007ae95e74f1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.343076] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-035381e0-17da-4a8d-af5d-d42aee500ce8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.357651] env[63372]: DEBUG nova.compute.provider_tree [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.464805] env[63372]: DEBUG nova.network.neutron [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.473585] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 558.615848] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "bd8cf6ba-8039-4c02-bda2-13c7d14f8903" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.616423] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "bd8cf6ba-8039-4c02-bda2-13c7d14f8903" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.736478] env[63372]: DEBUG nova.network.neutron [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.865459] env[63372]: DEBUG nova.scheduler.client.report [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 559.240805] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Releasing lock "refresh_cache-292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.241262] env[63372]: DEBUG nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 559.241338] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 559.241975] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33df26e0-ae5a-49de-b578-fe48d8f797e8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.257587] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b29dbb-158e-4af9-8114-a6233a7cb173 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.279567] env[63372]: ERROR nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1e3d9db7-b8f0-4221-9a0f-8538208cb01b, please check neutron logs for more information. [ 559.279567] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 559.279567] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.279567] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 559.279567] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.279567] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 559.279567] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.279567] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 559.279567] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.279567] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 559.279567] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.279567] env[63372]: ERROR nova.compute.manager raise self.value [ 559.279567] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.279567] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 559.279567] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.279567] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 559.280161] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.280161] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 559.280161] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1e3d9db7-b8f0-4221-9a0f-8538208cb01b, please check neutron logs for more information. [ 559.280161] env[63372]: ERROR nova.compute.manager [ 559.280161] env[63372]: Traceback (most recent call last): [ 559.280161] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 559.280161] env[63372]: listener.cb(fileno) [ 559.280161] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.280161] env[63372]: result = function(*args, **kwargs) [ 559.280161] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.280161] env[63372]: return func(*args, **kwargs) [ 559.280161] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.280161] env[63372]: raise e [ 559.280161] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.280161] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 559.280161] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.280161] env[63372]: created_port_ids = self._update_ports_for_instance( [ 559.280161] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.280161] env[63372]: with excutils.save_and_reraise_exception(): [ 559.280161] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.280161] env[63372]: self.force_reraise() [ 559.280161] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.280161] env[63372]: raise self.value [ 559.280161] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.280161] env[63372]: updated_port = self._update_port( [ 559.280161] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.280161] env[63372]: _ensure_no_port_binding_failure(port) [ 559.280161] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.280161] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 559.280910] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 1e3d9db7-b8f0-4221-9a0f-8538208cb01b, please check neutron logs for more information. [ 559.280910] env[63372]: Removing descriptor: 14 [ 559.280910] env[63372]: ERROR nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1e3d9db7-b8f0-4221-9a0f-8538208cb01b, please check neutron logs for more information. [ 559.280910] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Traceback (most recent call last): [ 559.280910] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 559.280910] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] yield resources [ 559.280910] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 559.280910] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] self.driver.spawn(context, instance, image_meta, [ 559.280910] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 559.280910] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.280910] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.280910] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] vm_ref = self.build_virtual_machine(instance, [ 559.281280] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.281280] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.281280] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.281280] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] for vif in network_info: [ 559.281280] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 559.281280] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] return self._sync_wrapper(fn, *args, **kwargs) [ 559.281280] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 559.281280] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] self.wait() [ 559.281280] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 559.281280] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] self[:] = self._gt.wait() [ 559.281280] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.281280] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] return self._exit_event.wait() [ 559.281280] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.285556] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] result = hub.switch() [ 559.285556] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.285556] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] return self.greenlet.switch() [ 559.285556] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.285556] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] result = function(*args, **kwargs) [ 559.285556] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.285556] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] return func(*args, **kwargs) [ 559.285556] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.285556] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] raise e [ 559.285556] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.285556] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] nwinfo = self.network_api.allocate_for_instance( [ 559.285556] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.285556] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] created_port_ids = self._update_ports_for_instance( [ 559.286118] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.286118] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] with excutils.save_and_reraise_exception(): [ 559.286118] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.286118] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] self.force_reraise() [ 559.286118] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.286118] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] raise self.value [ 559.286118] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.286118] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] updated_port = self._update_port( [ 559.286118] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.286118] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] _ensure_no_port_binding_failure(port) [ 559.286118] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.286118] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] raise exception.PortBindingFailed(port_id=port['id']) [ 559.286527] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] nova.exception.PortBindingFailed: Binding failed for port 1e3d9db7-b8f0-4221-9a0f-8538208cb01b, please check neutron logs for more information. [ 559.286527] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] [ 559.286527] env[63372]: INFO nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Terminating instance [ 559.287751] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44 could not be found. [ 559.287948] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 559.288141] env[63372]: INFO nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Took 0.05 seconds to destroy the instance on the hypervisor. [ 559.288383] env[63372]: DEBUG oslo.service.loopingcall [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.289252] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "refresh_cache-ee2f5088-616b-4237-93c9-fac5e0d6aba2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.289252] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquired lock "refresh_cache-ee2f5088-616b-4237-93c9-fac5e0d6aba2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.289252] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 559.292431] env[63372]: DEBUG nova.compute.manager [-] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.292431] env[63372]: DEBUG nova.network.neutron [-] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 559.334459] env[63372]: DEBUG nova.network.neutron [-] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.374659] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.918s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.374659] env[63372]: ERROR nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 69372e7a-1cad-4678-8591-897c91ccc7e1, please check neutron logs for more information. [ 559.374659] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Traceback (most recent call last): [ 559.374659] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 559.374659] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] self.driver.spawn(context, instance, image_meta, [ 559.374659] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 559.374659] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.374659] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.374659] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] vm_ref = self.build_virtual_machine(instance, [ 559.375048] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.375048] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.375048] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.375048] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] for vif in network_info: [ 559.375048] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 559.375048] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] return self._sync_wrapper(fn, *args, **kwargs) [ 559.375048] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 559.375048] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] self.wait() [ 559.375048] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 559.375048] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] self[:] = self._gt.wait() [ 559.375048] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.375048] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] return self._exit_event.wait() [ 559.375048] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.375394] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] result = hub.switch() [ 559.375394] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.375394] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] return self.greenlet.switch() [ 559.375394] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.375394] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] result = function(*args, **kwargs) [ 559.375394] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.375394] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] return func(*args, **kwargs) [ 559.375394] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.375394] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] raise e [ 559.375394] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.375394] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] nwinfo = self.network_api.allocate_for_instance( [ 559.375394] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.375394] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] created_port_ids = self._update_ports_for_instance( [ 559.375736] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.375736] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] with excutils.save_and_reraise_exception(): [ 559.375736] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.375736] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] self.force_reraise() [ 559.375736] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.375736] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] raise self.value [ 559.375736] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.375736] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] updated_port = self._update_port( [ 559.375736] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.375736] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] _ensure_no_port_binding_failure(port) [ 559.375736] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.375736] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] raise exception.PortBindingFailed(port_id=port['id']) [ 559.376070] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] nova.exception.PortBindingFailed: Binding failed for port 69372e7a-1cad-4678-8591-897c91ccc7e1, please check neutron logs for more information. [ 559.376070] env[63372]: ERROR nova.compute.manager [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] [ 559.376070] env[63372]: DEBUG nova.compute.utils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Binding failed for port 69372e7a-1cad-4678-8591-897c91ccc7e1, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 559.376541] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.305s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.378596] env[63372]: INFO nova.compute.claims [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 559.384083] env[63372]: DEBUG nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Build of instance 91351ef7-a282-4f57-b96b-ba0af6d0e074 was re-scheduled: Binding failed for port 69372e7a-1cad-4678-8591-897c91ccc7e1, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 559.384083] env[63372]: DEBUG nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 559.384083] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Acquiring lock "refresh_cache-91351ef7-a282-4f57-b96b-ba0af6d0e074" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.384083] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Acquired lock "refresh_cache-91351ef7-a282-4f57-b96b-ba0af6d0e074" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.384083] env[63372]: DEBUG nova.network.neutron [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 559.487278] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 559.521316] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 559.521555] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 559.521733] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 559.521941] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 559.522149] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 559.522412] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 559.522568] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 559.522759] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 559.522943] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 559.523119] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 559.523498] env[63372]: DEBUG nova.virt.hardware [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 559.524388] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20bf5cf-3df0-4384-9667-109ce113f380 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.535078] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b041dc6b-10db-4bb9-b29d-68775234ee5e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.559908] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Successfully created port: 9eb8da80-70b2-45fe-92f5-022d9b9a217c {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 559.815146] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.838764] env[63372]: DEBUG nova.network.neutron [-] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.927330] env[63372]: DEBUG nova.network.neutron [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.980921] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.085268] env[63372]: DEBUG nova.compute.manager [req-0c04c336-cf7f-4de6-b55e-31b849bc9378 req-0bafc71f-46cc-442c-b81a-60145bf915f6 service nova] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Received event network-vif-deleted-d0f239ea-5afd-491e-a640-c038817aee6a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.085268] env[63372]: DEBUG nova.compute.manager [req-0c04c336-cf7f-4de6-b55e-31b849bc9378 req-0bafc71f-46cc-442c-b81a-60145bf915f6 service nova] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Received event network-changed-2bdb48e9-8cdf-4e6d-b8ca-12d83d6a2b26 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.085268] env[63372]: DEBUG nova.compute.manager [req-0c04c336-cf7f-4de6-b55e-31b849bc9378 req-0bafc71f-46cc-442c-b81a-60145bf915f6 service nova] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Refreshing instance network info cache due to event network-changed-2bdb48e9-8cdf-4e6d-b8ca-12d83d6a2b26. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 560.085268] env[63372]: DEBUG oslo_concurrency.lockutils [req-0c04c336-cf7f-4de6-b55e-31b849bc9378 req-0bafc71f-46cc-442c-b81a-60145bf915f6 service nova] Acquiring lock "refresh_cache-292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.085268] env[63372]: DEBUG oslo_concurrency.lockutils [req-0c04c336-cf7f-4de6-b55e-31b849bc9378 req-0bafc71f-46cc-442c-b81a-60145bf915f6 service nova] Acquired lock "refresh_cache-292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.085779] env[63372]: DEBUG nova.network.neutron [req-0c04c336-cf7f-4de6-b55e-31b849bc9378 req-0bafc71f-46cc-442c-b81a-60145bf915f6 service nova] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Refreshing network info cache for port 2bdb48e9-8cdf-4e6d-b8ca-12d83d6a2b26 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 560.110360] env[63372]: DEBUG nova.network.neutron [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.344520] env[63372]: INFO nova.compute.manager [-] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Took 1.05 seconds to deallocate network for instance. [ 560.349428] env[63372]: DEBUG nova.compute.claims [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 560.350257] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.486871] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Releasing lock "refresh_cache-ee2f5088-616b-4237-93c9-fac5e0d6aba2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.487599] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 560.487599] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 560.488093] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5c46743f-7508-40f5-bb1e-98af4686a337 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.497063] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b22a8f9-9478-49be-ad4f-d7eed11a0265 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.518749] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ee2f5088-616b-4237-93c9-fac5e0d6aba2 could not be found. [ 560.519036] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 560.519280] env[63372]: INFO nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 560.519630] env[63372]: DEBUG oslo.service.loopingcall [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 560.519940] env[63372]: DEBUG nova.compute.manager [-] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 560.520083] env[63372]: DEBUG nova.network.neutron [-] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 560.541178] env[63372]: DEBUG nova.network.neutron [-] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.605331] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Acquiring lock "6aa468f8-cd63-459b-8302-8bc22580b01d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.605540] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Lock "6aa468f8-cd63-459b-8302-8bc22580b01d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.612571] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Releasing lock "refresh_cache-91351ef7-a282-4f57-b96b-ba0af6d0e074" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.612780] env[63372]: DEBUG nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 560.612950] env[63372]: DEBUG nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 560.613126] env[63372]: DEBUG nova.network.neutron [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 560.649777] env[63372]: DEBUG nova.network.neutron [req-0c04c336-cf7f-4de6-b55e-31b849bc9378 req-0bafc71f-46cc-442c-b81a-60145bf915f6 service nova] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.659913] env[63372]: DEBUG nova.network.neutron [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.692804] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Acquiring lock "ae86a977-0a9a-46e5-9072-cfb772324393" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.692804] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Lock "ae86a977-0a9a-46e5-9072-cfb772324393" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.763975] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918b0221-629f-47ca-afeb-1488bde29edd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.777011] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-110a9098-c4e7-4af7-a719-70ff84109ad1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.812152] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf2f966-4934-4612-8efe-5285a49da282 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.820055] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10695692-434b-438c-834b-220f40f766a2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.833407] env[63372]: DEBUG nova.compute.provider_tree [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.959800] env[63372]: DEBUG nova.network.neutron [req-0c04c336-cf7f-4de6-b55e-31b849bc9378 req-0bafc71f-46cc-442c-b81a-60145bf915f6 service nova] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.043784] env[63372]: DEBUG nova.network.neutron [-] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.163274] env[63372]: DEBUG nova.network.neutron [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.336897] env[63372]: DEBUG nova.scheduler.client.report [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 561.464723] env[63372]: DEBUG oslo_concurrency.lockutils [req-0c04c336-cf7f-4de6-b55e-31b849bc9378 req-0bafc71f-46cc-442c-b81a-60145bf915f6 service nova] Releasing lock "refresh_cache-292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.547686] env[63372]: INFO nova.compute.manager [-] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Took 1.03 seconds to deallocate network for instance. [ 561.558400] env[63372]: DEBUG nova.compute.claims [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 561.558923] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.575423] env[63372]: DEBUG nova.compute.manager [req-18f69fa6-d8d7-4911-89e3-6ebc32c42cc9 req-13642990-61e6-4bf9-95f5-72d09fc53966 service nova] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Received event network-changed-1e3d9db7-b8f0-4221-9a0f-8538208cb01b {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 561.575508] env[63372]: DEBUG nova.compute.manager [req-18f69fa6-d8d7-4911-89e3-6ebc32c42cc9 req-13642990-61e6-4bf9-95f5-72d09fc53966 service nova] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Refreshing instance network info cache due to event network-changed-1e3d9db7-b8f0-4221-9a0f-8538208cb01b. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 561.575766] env[63372]: DEBUG oslo_concurrency.lockutils [req-18f69fa6-d8d7-4911-89e3-6ebc32c42cc9 req-13642990-61e6-4bf9-95f5-72d09fc53966 service nova] Acquiring lock "refresh_cache-ee2f5088-616b-4237-93c9-fac5e0d6aba2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.576303] env[63372]: DEBUG oslo_concurrency.lockutils [req-18f69fa6-d8d7-4911-89e3-6ebc32c42cc9 req-13642990-61e6-4bf9-95f5-72d09fc53966 service nova] Acquired lock "refresh_cache-ee2f5088-616b-4237-93c9-fac5e0d6aba2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.576303] env[63372]: DEBUG nova.network.neutron [req-18f69fa6-d8d7-4911-89e3-6ebc32c42cc9 req-13642990-61e6-4bf9-95f5-72d09fc53966 service nova] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Refreshing network info cache for port 1e3d9db7-b8f0-4221-9a0f-8538208cb01b {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 561.668141] env[63372]: INFO nova.compute.manager [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] [instance: 91351ef7-a282-4f57-b96b-ba0af6d0e074] Took 1.05 seconds to deallocate network for instance. [ 561.672165] env[63372]: ERROR nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 66605f0f-5cf3-4815-9309-0d3bcd4144d8, please check neutron logs for more information. [ 561.672165] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 561.672165] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.672165] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 561.672165] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.672165] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 561.672165] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.672165] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 561.672165] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.672165] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 561.672165] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.672165] env[63372]: ERROR nova.compute.manager raise self.value [ 561.672165] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.672165] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 561.672165] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.672165] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 561.672710] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.672710] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 561.672710] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 66605f0f-5cf3-4815-9309-0d3bcd4144d8, please check neutron logs for more information. [ 561.672710] env[63372]: ERROR nova.compute.manager [ 561.672710] env[63372]: Traceback (most recent call last): [ 561.672710] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 561.672710] env[63372]: listener.cb(fileno) [ 561.672710] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.672710] env[63372]: result = function(*args, **kwargs) [ 561.672710] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.672710] env[63372]: return func(*args, **kwargs) [ 561.672710] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.672710] env[63372]: raise e [ 561.672710] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.672710] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 561.672710] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.672710] env[63372]: created_port_ids = self._update_ports_for_instance( [ 561.672710] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.672710] env[63372]: with excutils.save_and_reraise_exception(): [ 561.672710] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.672710] env[63372]: self.force_reraise() [ 561.672710] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.672710] env[63372]: raise self.value [ 561.672710] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.672710] env[63372]: updated_port = self._update_port( [ 561.672710] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.672710] env[63372]: _ensure_no_port_binding_failure(port) [ 561.672710] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.672710] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 561.674304] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 66605f0f-5cf3-4815-9309-0d3bcd4144d8, please check neutron logs for more information. [ 561.674304] env[63372]: Removing descriptor: 18 [ 561.674304] env[63372]: ERROR nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 66605f0f-5cf3-4815-9309-0d3bcd4144d8, please check neutron logs for more information. [ 561.674304] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Traceback (most recent call last): [ 561.674304] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 561.674304] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] yield resources [ 561.674304] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 561.674304] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] self.driver.spawn(context, instance, image_meta, [ 561.674304] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 561.674304] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 561.674304] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 561.674304] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] vm_ref = self.build_virtual_machine(instance, [ 561.674915] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 561.674915] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] vif_infos = vmwarevif.get_vif_info(self._session, [ 561.674915] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 561.674915] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] for vif in network_info: [ 561.674915] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 561.674915] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] return self._sync_wrapper(fn, *args, **kwargs) [ 561.674915] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 561.674915] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] self.wait() [ 561.674915] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 561.674915] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] self[:] = self._gt.wait() [ 561.674915] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 561.674915] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] return self._exit_event.wait() [ 561.674915] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 561.676837] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] result = hub.switch() [ 561.676837] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 561.676837] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] return self.greenlet.switch() [ 561.676837] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.676837] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] result = function(*args, **kwargs) [ 561.676837] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.676837] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] return func(*args, **kwargs) [ 561.676837] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.676837] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] raise e [ 561.676837] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.676837] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] nwinfo = self.network_api.allocate_for_instance( [ 561.676837] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.676837] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] created_port_ids = self._update_ports_for_instance( [ 561.677566] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.677566] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] with excutils.save_and_reraise_exception(): [ 561.677566] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.677566] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] self.force_reraise() [ 561.677566] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.677566] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] raise self.value [ 561.677566] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.677566] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] updated_port = self._update_port( [ 561.677566] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.677566] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] _ensure_no_port_binding_failure(port) [ 561.677566] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.677566] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] raise exception.PortBindingFailed(port_id=port['id']) [ 561.677926] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] nova.exception.PortBindingFailed: Binding failed for port 66605f0f-5cf3-4815-9309-0d3bcd4144d8, please check neutron logs for more information. [ 561.677926] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] [ 561.677926] env[63372]: INFO nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Terminating instance [ 561.677926] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "refresh_cache-39e22917-3ea8-457c-9d6e-679a848c3fd3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.677926] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquired lock "refresh_cache-39e22917-3ea8-457c-9d6e-679a848c3fd3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.677926] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 561.844242] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.844593] env[63372]: DEBUG nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 561.852070] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.153s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.854848] env[63372]: INFO nova.compute.claims [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 561.864373] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.864608] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.146123] env[63372]: DEBUG nova.network.neutron [req-18f69fa6-d8d7-4911-89e3-6ebc32c42cc9 req-13642990-61e6-4bf9-95f5-72d09fc53966 service nova] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.292981] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.356123] env[63372]: DEBUG nova.compute.utils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 562.357522] env[63372]: DEBUG nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 562.359977] env[63372]: DEBUG nova.network.neutron [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 562.446345] env[63372]: DEBUG nova.network.neutron [req-18f69fa6-d8d7-4911-89e3-6ebc32c42cc9 req-13642990-61e6-4bf9-95f5-72d09fc53966 service nova] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.504966] env[63372]: DEBUG nova.policy [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fcf8717bfe6b4492a7f64ea3624bcce8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2c98aec976634addb624aa1b945c6beb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 562.556110] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.578666] env[63372]: ERROR nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9eb8da80-70b2-45fe-92f5-022d9b9a217c, please check neutron logs for more information. [ 562.578666] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 562.578666] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.578666] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 562.578666] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 562.578666] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 562.578666] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 562.578666] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 562.578666] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.578666] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 562.578666] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.578666] env[63372]: ERROR nova.compute.manager raise self.value [ 562.578666] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 562.578666] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 562.578666] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.578666] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 562.579136] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.579136] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 562.579136] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9eb8da80-70b2-45fe-92f5-022d9b9a217c, please check neutron logs for more information. [ 562.579136] env[63372]: ERROR nova.compute.manager [ 562.579136] env[63372]: Traceback (most recent call last): [ 562.579136] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 562.579136] env[63372]: listener.cb(fileno) [ 562.579136] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.579136] env[63372]: result = function(*args, **kwargs) [ 562.579136] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 562.579136] env[63372]: return func(*args, **kwargs) [ 562.579136] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 562.579136] env[63372]: raise e [ 562.579136] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.579136] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 562.579136] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 562.579136] env[63372]: created_port_ids = self._update_ports_for_instance( [ 562.579136] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 562.579136] env[63372]: with excutils.save_and_reraise_exception(): [ 562.579136] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.579136] env[63372]: self.force_reraise() [ 562.579136] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.579136] env[63372]: raise self.value [ 562.579136] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 562.579136] env[63372]: updated_port = self._update_port( [ 562.579136] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.579136] env[63372]: _ensure_no_port_binding_failure(port) [ 562.579136] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.579136] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 562.579986] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 9eb8da80-70b2-45fe-92f5-022d9b9a217c, please check neutron logs for more information. [ 562.579986] env[63372]: Removing descriptor: 16 [ 562.579986] env[63372]: ERROR nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9eb8da80-70b2-45fe-92f5-022d9b9a217c, please check neutron logs for more information. [ 562.579986] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Traceback (most recent call last): [ 562.579986] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 562.579986] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] yield resources [ 562.579986] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 562.579986] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] self.driver.spawn(context, instance, image_meta, [ 562.579986] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 562.579986] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] self._vmops.spawn(context, instance, image_meta, injected_files, [ 562.579986] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 562.579986] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] vm_ref = self.build_virtual_machine(instance, [ 562.580424] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 562.580424] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] vif_infos = vmwarevif.get_vif_info(self._session, [ 562.580424] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 562.580424] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] for vif in network_info: [ 562.580424] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 562.580424] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] return self._sync_wrapper(fn, *args, **kwargs) [ 562.580424] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 562.580424] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] self.wait() [ 562.580424] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 562.580424] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] self[:] = self._gt.wait() [ 562.580424] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 562.580424] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] return self._exit_event.wait() [ 562.580424] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 562.580762] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] result = hub.switch() [ 562.580762] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 562.580762] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] return self.greenlet.switch() [ 562.580762] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.580762] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] result = function(*args, **kwargs) [ 562.580762] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 562.580762] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] return func(*args, **kwargs) [ 562.580762] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 562.580762] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] raise e [ 562.580762] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.580762] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] nwinfo = self.network_api.allocate_for_instance( [ 562.580762] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 562.580762] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] created_port_ids = self._update_ports_for_instance( [ 562.581164] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 562.581164] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] with excutils.save_and_reraise_exception(): [ 562.581164] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.581164] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] self.force_reraise() [ 562.581164] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.581164] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] raise self.value [ 562.581164] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 562.581164] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] updated_port = self._update_port( [ 562.581164] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.581164] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] _ensure_no_port_binding_failure(port) [ 562.581164] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.581164] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] raise exception.PortBindingFailed(port_id=port['id']) [ 562.581466] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] nova.exception.PortBindingFailed: Binding failed for port 9eb8da80-70b2-45fe-92f5-022d9b9a217c, please check neutron logs for more information. [ 562.581466] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] [ 562.581466] env[63372]: INFO nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Terminating instance [ 562.583269] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "refresh_cache-7803b83f-2944-4e3b-87b1-0d3cea3455ad" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.583486] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquired lock "refresh_cache-7803b83f-2944-4e3b-87b1-0d3cea3455ad" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.583686] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 562.589539] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Acquiring lock "779e3559-37a5-49b2-b17a-bf95d0d8d23c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.589810] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Lock "779e3559-37a5-49b2-b17a-bf95d0d8d23c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.730712] env[63372]: INFO nova.scheduler.client.report [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Deleted allocations for instance 91351ef7-a282-4f57-b96b-ba0af6d0e074 [ 562.864340] env[63372]: DEBUG nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 562.949607] env[63372]: DEBUG oslo_concurrency.lockutils [req-18f69fa6-d8d7-4911-89e3-6ebc32c42cc9 req-13642990-61e6-4bf9-95f5-72d09fc53966 service nova] Releasing lock "refresh_cache-ee2f5088-616b-4237-93c9-fac5e0d6aba2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.061124] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Releasing lock "refresh_cache-39e22917-3ea8-457c-9d6e-679a848c3fd3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.061805] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 563.061805] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 563.062484] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-332d021f-8c26-4214-b607-f092430f0ab0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.072360] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd381ff-74a8-4dd3-b2f0-668755162777 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.106014] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 39e22917-3ea8-457c-9d6e-679a848c3fd3 could not be found. [ 563.106917] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 563.106917] env[63372]: INFO nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 563.106917] env[63372]: DEBUG oslo.service.loopingcall [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 563.111217] env[63372]: DEBUG nova.compute.manager [-] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 563.111217] env[63372]: DEBUG nova.network.neutron [-] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 563.120413] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 563.153447] env[63372]: DEBUG nova.network.neutron [-] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 563.241401] env[63372]: DEBUG oslo_concurrency.lockutils [None req-218666db-a90f-4406-8e5b-b62ee42d5927 tempest-ServersTestJSON-534270014 tempest-ServersTestJSON-534270014-project-member] Lock "91351ef7-a282-4f57-b96b-ba0af6d0e074" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.729s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.300037] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846880ef-d415-4e5c-9ef8-c8f27de9bd76 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.314727] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18df32c2-77ae-4459-bd07-415a0e6ef05e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.349884] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358d14cd-bdfa-4395-af4a-93cb186c10a9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.358971] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc91929-6714-4246-a960-6b96fd40e652 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.385234] env[63372]: DEBUG nova.compute.provider_tree [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.426692] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.432520] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Acquiring lock "cf673ac1-2c7d-468b-83ec-c723d5182457" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.432520] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Lock "cf673ac1-2c7d-468b-83ec-c723d5182457" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.513015] env[63372]: DEBUG nova.compute.manager [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Received event network-vif-deleted-2bdb48e9-8cdf-4e6d-b8ca-12d83d6a2b26 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 563.514371] env[63372]: DEBUG nova.compute.manager [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Received event network-changed-66605f0f-5cf3-4815-9309-0d3bcd4144d8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 563.514371] env[63372]: DEBUG nova.compute.manager [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Refreshing instance network info cache due to event network-changed-66605f0f-5cf3-4815-9309-0d3bcd4144d8. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 563.514371] env[63372]: DEBUG oslo_concurrency.lockutils [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] Acquiring lock "refresh_cache-39e22917-3ea8-457c-9d6e-679a848c3fd3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.514371] env[63372]: DEBUG oslo_concurrency.lockutils [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] Acquired lock "refresh_cache-39e22917-3ea8-457c-9d6e-679a848c3fd3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.514371] env[63372]: DEBUG nova.network.neutron [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Refreshing network info cache for port 66605f0f-5cf3-4815-9309-0d3bcd4144d8 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 563.622897] env[63372]: DEBUG nova.network.neutron [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Successfully created port: 88693cd4-b712-426d-8db2-67c49dc783d5 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 563.657844] env[63372]: DEBUG nova.network.neutron [-] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.749639] env[63372]: DEBUG nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 563.888912] env[63372]: DEBUG nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 563.891723] env[63372]: DEBUG nova.scheduler.client.report [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 563.928540] env[63372]: DEBUG nova.virt.hardware [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 563.928787] env[63372]: DEBUG nova.virt.hardware [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 563.928941] env[63372]: DEBUG nova.virt.hardware [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 563.929285] env[63372]: DEBUG nova.virt.hardware [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 563.929674] env[63372]: DEBUG nova.virt.hardware [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 563.929674] env[63372]: DEBUG nova.virt.hardware [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 563.929862] env[63372]: DEBUG nova.virt.hardware [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 563.930020] env[63372]: DEBUG nova.virt.hardware [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 563.930109] env[63372]: DEBUG nova.virt.hardware [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 563.930432] env[63372]: DEBUG nova.virt.hardware [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 563.930432] env[63372]: DEBUG nova.virt.hardware [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 563.931294] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b276f11-c76c-4945-854f-a3c97326d81b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.934154] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Releasing lock "refresh_cache-7803b83f-2944-4e3b-87b1-0d3cea3455ad" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.934557] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 563.934832] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 563.937482] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c993d97-7783-4608-acec-20c425e4dd9c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.944862] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d2c4ec-f396-43bc-be49-1277e7b65a51 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.964652] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9ce275-8475-4b3a-8ea3-50a93cb86c4b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.986480] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7803b83f-2944-4e3b-87b1-0d3cea3455ad could not be found. [ 563.986691] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 563.986869] env[63372]: INFO nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Took 0.05 seconds to destroy the instance on the hypervisor. [ 563.987145] env[63372]: DEBUG oslo.service.loopingcall [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 563.987388] env[63372]: DEBUG nova.compute.manager [-] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 563.987461] env[63372]: DEBUG nova.network.neutron [-] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 564.036083] env[63372]: DEBUG nova.network.neutron [-] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 564.053465] env[63372]: DEBUG nova.network.neutron [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 564.158621] env[63372]: INFO nova.compute.manager [-] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Took 1.05 seconds to deallocate network for instance. [ 564.161961] env[63372]: DEBUG nova.compute.claims [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 564.162090] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.231106] env[63372]: DEBUG nova.compute.manager [req-cb2b54fb-fc79-4c26-8af4-afa75d66c3e4 req-f26914ea-20d4-4963-9794-a39acd68d486 service nova] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Received event network-vif-deleted-1e3d9db7-b8f0-4221-9a0f-8538208cb01b {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 564.291690] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.399369] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.399503] env[63372]: DEBUG nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 564.405803] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.421s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.408136] env[63372]: INFO nova.compute.claims [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 564.527427] env[63372]: DEBUG nova.network.neutron [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.539649] env[63372]: DEBUG nova.network.neutron [-] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.908974] env[63372]: DEBUG nova.compute.utils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 564.910226] env[63372]: DEBUG nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 564.914078] env[63372]: DEBUG nova.network.neutron [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 565.031713] env[63372]: DEBUG oslo_concurrency.lockutils [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] Releasing lock "refresh_cache-39e22917-3ea8-457c-9d6e-679a848c3fd3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.031713] env[63372]: DEBUG nova.compute.manager [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Received event network-vif-deleted-66605f0f-5cf3-4815-9309-0d3bcd4144d8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 565.032037] env[63372]: DEBUG nova.compute.manager [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Received event network-changed-9eb8da80-70b2-45fe-92f5-022d9b9a217c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 565.032120] env[63372]: DEBUG nova.compute.manager [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Refreshing instance network info cache due to event network-changed-9eb8da80-70b2-45fe-92f5-022d9b9a217c. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 565.032297] env[63372]: DEBUG oslo_concurrency.lockutils [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] Acquiring lock "refresh_cache-7803b83f-2944-4e3b-87b1-0d3cea3455ad" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.032732] env[63372]: DEBUG oslo_concurrency.lockutils [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] Acquired lock "refresh_cache-7803b83f-2944-4e3b-87b1-0d3cea3455ad" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.032732] env[63372]: DEBUG nova.network.neutron [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Refreshing network info cache for port 9eb8da80-70b2-45fe-92f5-022d9b9a217c {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 565.043632] env[63372]: INFO nova.compute.manager [-] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Took 1.06 seconds to deallocate network for instance. [ 565.045642] env[63372]: DEBUG nova.compute.claims [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 565.045887] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.079132] env[63372]: DEBUG nova.policy [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fcf8717bfe6b4492a7f64ea3624bcce8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2c98aec976634addb624aa1b945c6beb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 565.419185] env[63372]: DEBUG nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 565.578040] env[63372]: DEBUG nova.network.neutron [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 565.832891] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22a73f9-bb20-49e3-a388-b7febdbb14b9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.843738] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce82345-6092-4a1d-9d33-33fca54e371a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.883587] env[63372]: DEBUG nova.network.neutron [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.885349] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c76b8e53-478b-417c-bbe3-3b19dbf29c8f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.899957] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d05827-19e4-4a12-9c1a-2bbd08fbc1e0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.917974] env[63372]: DEBUG nova.compute.provider_tree [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.958486] env[63372]: DEBUG nova.network.neutron [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Successfully created port: 32a3c00b-7133-4607-b512-85a673630d43 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 566.396055] env[63372]: DEBUG oslo_concurrency.lockutils [req-1b6f1c8c-dbda-4363-8917-5e7f3d934b8a req-b732b4ac-dd5f-4c53-8e29-91925d78220c service nova] Releasing lock "refresh_cache-7803b83f-2944-4e3b-87b1-0d3cea3455ad" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.420945] env[63372]: DEBUG nova.scheduler.client.report [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.435901] env[63372]: DEBUG nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 566.477133] env[63372]: DEBUG nova.virt.hardware [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 566.478711] env[63372]: DEBUG nova.virt.hardware [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 566.478711] env[63372]: DEBUG nova.virt.hardware [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 566.478711] env[63372]: DEBUG nova.virt.hardware [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 566.478711] env[63372]: DEBUG nova.virt.hardware [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 566.478711] env[63372]: DEBUG nova.virt.hardware [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 566.478952] env[63372]: DEBUG nova.virt.hardware [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 566.478952] env[63372]: DEBUG nova.virt.hardware [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 566.478952] env[63372]: DEBUG nova.virt.hardware [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 566.479078] env[63372]: DEBUG nova.virt.hardware [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 566.479245] env[63372]: DEBUG nova.virt.hardware [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 566.480145] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01267ba5-c04b-4f4f-b609-4e27d93c2f0b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.489480] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5627be6b-2f5c-4e84-a400-e4fae3090b6d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.848598] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.848894] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.927049] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.927567] env[63372]: DEBUG nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 566.931658] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.968s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.935132] env[63372]: INFO nova.compute.claims [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 566.990119] env[63372]: ERROR nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 88693cd4-b712-426d-8db2-67c49dc783d5, please check neutron logs for more information. [ 566.990119] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 566.990119] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.990119] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 566.990119] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 566.990119] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 566.990119] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 566.990119] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 566.990119] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.990119] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 566.990119] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.990119] env[63372]: ERROR nova.compute.manager raise self.value [ 566.990119] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 566.990119] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 566.990119] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.990119] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 566.990596] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.990596] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 566.990596] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 88693cd4-b712-426d-8db2-67c49dc783d5, please check neutron logs for more information. [ 566.990596] env[63372]: ERROR nova.compute.manager [ 566.990596] env[63372]: Traceback (most recent call last): [ 566.990596] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 566.990596] env[63372]: listener.cb(fileno) [ 566.990596] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.990596] env[63372]: result = function(*args, **kwargs) [ 566.990596] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 566.990596] env[63372]: return func(*args, **kwargs) [ 566.990596] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.990596] env[63372]: raise e [ 566.990596] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.990596] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 566.990596] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 566.990596] env[63372]: created_port_ids = self._update_ports_for_instance( [ 566.990596] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 566.990596] env[63372]: with excutils.save_and_reraise_exception(): [ 566.990596] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.990596] env[63372]: self.force_reraise() [ 566.990596] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.990596] env[63372]: raise self.value [ 566.990596] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 566.990596] env[63372]: updated_port = self._update_port( [ 566.990596] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.990596] env[63372]: _ensure_no_port_binding_failure(port) [ 566.990596] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.990596] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 566.991409] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 88693cd4-b712-426d-8db2-67c49dc783d5, please check neutron logs for more information. [ 566.991409] env[63372]: Removing descriptor: 14 [ 566.991409] env[63372]: ERROR nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 88693cd4-b712-426d-8db2-67c49dc783d5, please check neutron logs for more information. [ 566.991409] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Traceback (most recent call last): [ 566.991409] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 566.991409] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] yield resources [ 566.991409] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 566.991409] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] self.driver.spawn(context, instance, image_meta, [ 566.991409] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 566.991409] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 566.991409] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 566.991409] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] vm_ref = self.build_virtual_machine(instance, [ 566.991820] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 566.991820] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] vif_infos = vmwarevif.get_vif_info(self._session, [ 566.991820] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 566.991820] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] for vif in network_info: [ 566.991820] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 566.991820] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] return self._sync_wrapper(fn, *args, **kwargs) [ 566.991820] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 566.991820] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] self.wait() [ 566.991820] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 566.991820] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] self[:] = self._gt.wait() [ 566.991820] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 566.991820] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] return self._exit_event.wait() [ 566.991820] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 566.992223] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] result = hub.switch() [ 566.992223] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 566.992223] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] return self.greenlet.switch() [ 566.992223] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.992223] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] result = function(*args, **kwargs) [ 566.992223] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 566.992223] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] return func(*args, **kwargs) [ 566.992223] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.992223] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] raise e [ 566.992223] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.992223] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] nwinfo = self.network_api.allocate_for_instance( [ 566.992223] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 566.992223] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] created_port_ids = self._update_ports_for_instance( [ 566.992586] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 566.992586] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] with excutils.save_and_reraise_exception(): [ 566.992586] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.992586] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] self.force_reraise() [ 566.992586] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.992586] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] raise self.value [ 566.992586] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 566.992586] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] updated_port = self._update_port( [ 566.992586] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.992586] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] _ensure_no_port_binding_failure(port) [ 566.992586] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.992586] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] raise exception.PortBindingFailed(port_id=port['id']) [ 566.993073] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] nova.exception.PortBindingFailed: Binding failed for port 88693cd4-b712-426d-8db2-67c49dc783d5, please check neutron logs for more information. [ 566.993073] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] [ 566.993073] env[63372]: INFO nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Terminating instance [ 566.996692] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Acquiring lock "refresh_cache-9ec742c1-894f-4baa-a451-01ebb4d2816a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.998351] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Acquired lock "refresh_cache-9ec742c1-894f-4baa-a451-01ebb4d2816a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.998351] env[63372]: DEBUG nova.network.neutron [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 567.011400] env[63372]: DEBUG nova.compute.manager [req-4787d73a-a6fd-4cc1-a9cd-d4f677b0848a req-6a3ceba3-bfff-4998-be2a-da89598f01ee service nova] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Received event network-vif-deleted-9eb8da80-70b2-45fe-92f5-022d9b9a217c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 567.437107] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Acquiring lock "9404a0f0-a3e5-43ee-bc20-e9566cfe5a60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.437381] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Lock "9404a0f0-a3e5-43ee-bc20-e9566cfe5a60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.444629] env[63372]: DEBUG nova.compute.utils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 567.447140] env[63372]: DEBUG nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 567.447140] env[63372]: DEBUG nova.network.neutron [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 567.532411] env[63372]: DEBUG nova.policy [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7fce5c89c32b4bd8bafcab78917302e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9fd44382e9a47bd9c202d75fac5497b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 567.550826] env[63372]: DEBUG nova.network.neutron [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.793642] env[63372]: DEBUG nova.compute.manager [req-7894a9ba-99ca-4ad6-b0c8-abe18307b5f5 req-e1762586-85c4-40e7-8430-c77baf38d7ec service nova] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Received event network-changed-88693cd4-b712-426d-8db2-67c49dc783d5 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 567.793642] env[63372]: DEBUG nova.compute.manager [req-7894a9ba-99ca-4ad6-b0c8-abe18307b5f5 req-e1762586-85c4-40e7-8430-c77baf38d7ec service nova] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Refreshing instance network info cache due to event network-changed-88693cd4-b712-426d-8db2-67c49dc783d5. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 567.793642] env[63372]: DEBUG oslo_concurrency.lockutils [req-7894a9ba-99ca-4ad6-b0c8-abe18307b5f5 req-e1762586-85c4-40e7-8430-c77baf38d7ec service nova] Acquiring lock "refresh_cache-9ec742c1-894f-4baa-a451-01ebb4d2816a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.955541] env[63372]: DEBUG nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 568.028036] env[63372]: DEBUG nova.network.neutron [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.321585] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959e4370-829c-454c-998f-3f9afdae2a17 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.329696] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f393d37-a717-4600-9c47-b566b9dcbc1c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.360787] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc09d60-121c-4240-83f9-615f1800c4c0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.368464] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be0f35f-5273-4952-b1fe-67334175cb84 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.381982] env[63372]: DEBUG nova.compute.provider_tree [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 568.469177] env[63372]: DEBUG nova.network.neutron [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Successfully created port: 92a840c0-9f37-4af2-9b74-f20b406f353d {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 568.536120] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Releasing lock "refresh_cache-9ec742c1-894f-4baa-a451-01ebb4d2816a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.536120] env[63372]: DEBUG nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 568.536120] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 568.536120] env[63372]: DEBUG oslo_concurrency.lockutils [req-7894a9ba-99ca-4ad6-b0c8-abe18307b5f5 req-e1762586-85c4-40e7-8430-c77baf38d7ec service nova] Acquired lock "refresh_cache-9ec742c1-894f-4baa-a451-01ebb4d2816a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.536120] env[63372]: DEBUG nova.network.neutron [req-7894a9ba-99ca-4ad6-b0c8-abe18307b5f5 req-e1762586-85c4-40e7-8430-c77baf38d7ec service nova] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Refreshing network info cache for port 88693cd4-b712-426d-8db2-67c49dc783d5 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 568.536334] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fa311a52-5ea2-4b33-bd69-2cf06898181d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.549740] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a783282c-8c19-4583-adb6-65e4f949a130 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.578835] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9ec742c1-894f-4baa-a451-01ebb4d2816a could not be found. [ 568.579082] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 568.579257] env[63372]: INFO nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Took 0.05 seconds to destroy the instance on the hypervisor. [ 568.579490] env[63372]: DEBUG oslo.service.loopingcall [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 568.579699] env[63372]: DEBUG nova.compute.manager [-] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 568.579788] env[63372]: DEBUG nova.network.neutron [-] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 568.626818] env[63372]: DEBUG nova.network.neutron [-] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.698671] env[63372]: ERROR nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 32a3c00b-7133-4607-b512-85a673630d43, please check neutron logs for more information. [ 568.698671] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 568.698671] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.698671] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 568.698671] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.698671] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 568.698671] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.698671] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 568.698671] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.698671] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 568.698671] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.698671] env[63372]: ERROR nova.compute.manager raise self.value [ 568.698671] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.698671] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 568.698671] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.698671] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 568.699572] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.699572] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 568.699572] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 32a3c00b-7133-4607-b512-85a673630d43, please check neutron logs for more information. [ 568.699572] env[63372]: ERROR nova.compute.manager [ 568.699572] env[63372]: Traceback (most recent call last): [ 568.699572] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 568.699572] env[63372]: listener.cb(fileno) [ 568.699572] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.699572] env[63372]: result = function(*args, **kwargs) [ 568.699572] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.699572] env[63372]: return func(*args, **kwargs) [ 568.699572] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.699572] env[63372]: raise e [ 568.699572] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.699572] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 568.699572] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.699572] env[63372]: created_port_ids = self._update_ports_for_instance( [ 568.699572] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.699572] env[63372]: with excutils.save_and_reraise_exception(): [ 568.699572] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.699572] env[63372]: self.force_reraise() [ 568.699572] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.699572] env[63372]: raise self.value [ 568.699572] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.699572] env[63372]: updated_port = self._update_port( [ 568.699572] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.699572] env[63372]: _ensure_no_port_binding_failure(port) [ 568.699572] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.699572] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 568.700718] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 32a3c00b-7133-4607-b512-85a673630d43, please check neutron logs for more information. [ 568.700718] env[63372]: Removing descriptor: 19 [ 568.700718] env[63372]: ERROR nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 32a3c00b-7133-4607-b512-85a673630d43, please check neutron logs for more information. [ 568.700718] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Traceback (most recent call last): [ 568.700718] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 568.700718] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] yield resources [ 568.700718] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.700718] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] self.driver.spawn(context, instance, image_meta, [ 568.700718] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 568.700718] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.700718] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.700718] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] vm_ref = self.build_virtual_machine(instance, [ 568.701069] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.701069] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.701069] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.701069] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] for vif in network_info: [ 568.701069] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.701069] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] return self._sync_wrapper(fn, *args, **kwargs) [ 568.701069] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.701069] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] self.wait() [ 568.701069] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.701069] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] self[:] = self._gt.wait() [ 568.701069] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.701069] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] return self._exit_event.wait() [ 568.701069] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.701538] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] result = hub.switch() [ 568.701538] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.701538] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] return self.greenlet.switch() [ 568.701538] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.701538] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] result = function(*args, **kwargs) [ 568.701538] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.701538] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] return func(*args, **kwargs) [ 568.701538] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.701538] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] raise e [ 568.701538] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.701538] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] nwinfo = self.network_api.allocate_for_instance( [ 568.701538] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.701538] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] created_port_ids = self._update_ports_for_instance( [ 568.701923] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.701923] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] with excutils.save_and_reraise_exception(): [ 568.701923] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.701923] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] self.force_reraise() [ 568.701923] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.701923] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] raise self.value [ 568.701923] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.701923] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] updated_port = self._update_port( [ 568.701923] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.701923] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] _ensure_no_port_binding_failure(port) [ 568.701923] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.701923] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] raise exception.PortBindingFailed(port_id=port['id']) [ 568.702256] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] nova.exception.PortBindingFailed: Binding failed for port 32a3c00b-7133-4607-b512-85a673630d43, please check neutron logs for more information. [ 568.702256] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] [ 568.702256] env[63372]: INFO nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Terminating instance [ 568.703497] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Acquiring lock "refresh_cache-dc9c4a56-e0af-473b-b47b-2061fe89c6d6" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.703705] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Acquired lock "refresh_cache-dc9c4a56-e0af-473b-b47b-2061fe89c6d6" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.703810] env[63372]: DEBUG nova.network.neutron [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 568.885346] env[63372]: DEBUG nova.scheduler.client.report [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 568.972771] env[63372]: DEBUG nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 569.002839] env[63372]: DEBUG nova.virt.hardware [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 569.003097] env[63372]: DEBUG nova.virt.hardware [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 569.003248] env[63372]: DEBUG nova.virt.hardware [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 569.003421] env[63372]: DEBUG nova.virt.hardware [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 569.003558] env[63372]: DEBUG nova.virt.hardware [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 569.003760] env[63372]: DEBUG nova.virt.hardware [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 569.003893] env[63372]: DEBUG nova.virt.hardware [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 569.005397] env[63372]: DEBUG nova.virt.hardware [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 569.005397] env[63372]: DEBUG nova.virt.hardware [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 569.005397] env[63372]: DEBUG nova.virt.hardware [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 569.005397] env[63372]: DEBUG nova.virt.hardware [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 569.005825] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b6ebf0-9eac-42b6-81e8-201f292ac46e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.014694] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7d7395-2016-424e-bd25-7a998a270d37 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.095782] env[63372]: DEBUG nova.network.neutron [req-7894a9ba-99ca-4ad6-b0c8-abe18307b5f5 req-e1762586-85c4-40e7-8430-c77baf38d7ec service nova] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.130848] env[63372]: DEBUG nova.network.neutron [-] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.244389] env[63372]: DEBUG nova.network.neutron [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.392307] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.392840] env[63372]: DEBUG nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 569.395430] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.256s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.479328] env[63372]: DEBUG nova.network.neutron [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.534838] env[63372]: DEBUG nova.network.neutron [req-7894a9ba-99ca-4ad6-b0c8-abe18307b5f5 req-e1762586-85c4-40e7-8430-c77baf38d7ec service nova] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.633620] env[63372]: INFO nova.compute.manager [-] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Took 1.05 seconds to deallocate network for instance. [ 569.636929] env[63372]: DEBUG nova.compute.claims [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 569.636929] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.901199] env[63372]: DEBUG nova.compute.utils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 569.906877] env[63372]: DEBUG nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 569.909025] env[63372]: DEBUG nova.network.neutron [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 569.983798] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Releasing lock "refresh_cache-dc9c4a56-e0af-473b-b47b-2061fe89c6d6" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.985544] env[63372]: DEBUG nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 569.985871] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 569.986919] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-73823a98-4286-48db-a6ad-60e990701417 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.998483] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77dce9b0-6d0d-4a17-805b-7398595c4880 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.028099] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dc9c4a56-e0af-473b-b47b-2061fe89c6d6 could not be found. [ 570.028684] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 570.028684] env[63372]: INFO nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 570.028763] env[63372]: DEBUG oslo.service.loopingcall [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 570.031620] env[63372]: DEBUG nova.compute.manager [-] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 570.031620] env[63372]: DEBUG nova.network.neutron [-] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 570.039772] env[63372]: DEBUG oslo_concurrency.lockutils [req-7894a9ba-99ca-4ad6-b0c8-abe18307b5f5 req-e1762586-85c4-40e7-8430-c77baf38d7ec service nova] Releasing lock "refresh_cache-9ec742c1-894f-4baa-a451-01ebb4d2816a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.069123] env[63372]: DEBUG nova.network.neutron [-] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.103029] env[63372]: DEBUG nova.policy [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '855ce9d887cd463cbf1a0f74e1197808', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b68c21179fb24008baf1da0d0119fb29', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 570.329137] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e0771b-3f75-4768-bfa1-d97587bec4f0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.338227] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad7e1d90-2180-46d0-ab1a-9c458686ce42 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.371192] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9bc5142-7916-41d4-951e-300ccaee3ebe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.379801] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159ddbbf-b862-447c-99c4-2883631038ac {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.394682] env[63372]: DEBUG nova.compute.provider_tree [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 570.406458] env[63372]: DEBUG nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 570.464857] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Acquiring lock "5adf6952-e019-4ea0-be91-8fe5fb83d53e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.464857] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Lock "5adf6952-e019-4ea0-be91-8fe5fb83d53e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.572276] env[63372]: DEBUG nova.network.neutron [-] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.636448] env[63372]: DEBUG nova.compute.manager [req-59c78604-a2a5-46f1-81e2-ae07652cbc87 req-0f39c39d-d091-4b32-8596-f81dbc9b9100 service nova] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Received event network-vif-deleted-88693cd4-b712-426d-8db2-67c49dc783d5 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 570.636697] env[63372]: DEBUG nova.compute.manager [req-59c78604-a2a5-46f1-81e2-ae07652cbc87 req-0f39c39d-d091-4b32-8596-f81dbc9b9100 service nova] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Received event network-changed-32a3c00b-7133-4607-b512-85a673630d43 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 570.636825] env[63372]: DEBUG nova.compute.manager [req-59c78604-a2a5-46f1-81e2-ae07652cbc87 req-0f39c39d-d091-4b32-8596-f81dbc9b9100 service nova] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Refreshing instance network info cache due to event network-changed-32a3c00b-7133-4607-b512-85a673630d43. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 570.637341] env[63372]: DEBUG oslo_concurrency.lockutils [req-59c78604-a2a5-46f1-81e2-ae07652cbc87 req-0f39c39d-d091-4b32-8596-f81dbc9b9100 service nova] Acquiring lock "refresh_cache-dc9c4a56-e0af-473b-b47b-2061fe89c6d6" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.637474] env[63372]: DEBUG oslo_concurrency.lockutils [req-59c78604-a2a5-46f1-81e2-ae07652cbc87 req-0f39c39d-d091-4b32-8596-f81dbc9b9100 service nova] Acquired lock "refresh_cache-dc9c4a56-e0af-473b-b47b-2061fe89c6d6" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.637713] env[63372]: DEBUG nova.network.neutron [req-59c78604-a2a5-46f1-81e2-ae07652cbc87 req-0f39c39d-d091-4b32-8596-f81dbc9b9100 service nova] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Refreshing network info cache for port 32a3c00b-7133-4607-b512-85a673630d43 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 570.901284] env[63372]: DEBUG nova.scheduler.client.report [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 571.076323] env[63372]: INFO nova.compute.manager [-] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Took 1.04 seconds to deallocate network for instance. [ 571.080542] env[63372]: DEBUG nova.compute.claims [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 571.080727] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.091643] env[63372]: DEBUG nova.network.neutron [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Successfully created port: 64490f1e-3e00-4fc9-92fb-a2e454998270 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 571.175601] env[63372]: DEBUG nova.network.neutron [req-59c78604-a2a5-46f1-81e2-ae07652cbc87 req-0f39c39d-d091-4b32-8596-f81dbc9b9100 service nova] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.411464] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.411464] env[63372]: ERROR nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d0f239ea-5afd-491e-a640-c038817aee6a, please check neutron logs for more information. [ 571.411464] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Traceback (most recent call last): [ 571.411464] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 571.411464] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] self.driver.spawn(context, instance, image_meta, [ 571.411464] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 571.411464] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 571.411464] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 571.411464] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] vm_ref = self.build_virtual_machine(instance, [ 571.411727] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 571.411727] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] vif_infos = vmwarevif.get_vif_info(self._session, [ 571.411727] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 571.411727] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] for vif in network_info: [ 571.411727] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 571.411727] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] return self._sync_wrapper(fn, *args, **kwargs) [ 571.411727] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 571.411727] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] self.wait() [ 571.411727] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 571.411727] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] self[:] = self._gt.wait() [ 571.411727] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 571.411727] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] return self._exit_event.wait() [ 571.411727] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 571.412082] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] result = hub.switch() [ 571.412082] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 571.412082] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] return self.greenlet.switch() [ 571.412082] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.412082] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] result = function(*args, **kwargs) [ 571.412082] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 571.412082] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] return func(*args, **kwargs) [ 571.412082] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 571.412082] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] raise e [ 571.412082] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.412082] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] nwinfo = self.network_api.allocate_for_instance( [ 571.412082] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 571.412082] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] created_port_ids = self._update_ports_for_instance( [ 571.412473] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 571.412473] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] with excutils.save_and_reraise_exception(): [ 571.412473] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.412473] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] self.force_reraise() [ 571.412473] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.412473] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] raise self.value [ 571.412473] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 571.412473] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] updated_port = self._update_port( [ 571.412473] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.412473] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] _ensure_no_port_binding_failure(port) [ 571.412473] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.412473] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] raise exception.PortBindingFailed(port_id=port['id']) [ 571.412826] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] nova.exception.PortBindingFailed: Binding failed for port d0f239ea-5afd-491e-a640-c038817aee6a, please check neutron logs for more information. [ 571.412826] env[63372]: ERROR nova.compute.manager [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] [ 571.412826] env[63372]: DEBUG nova.compute.utils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Binding failed for port d0f239ea-5afd-491e-a640-c038817aee6a, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 571.412826] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.061s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.416404] env[63372]: DEBUG nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Build of instance d290c8b5-422b-43c6-825e-52ea98516e6c was re-scheduled: Binding failed for port d0f239ea-5afd-491e-a640-c038817aee6a, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 571.416991] env[63372]: DEBUG nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 571.417424] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Acquiring lock "refresh_cache-d290c8b5-422b-43c6-825e-52ea98516e6c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.417684] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Acquired lock "refresh_cache-d290c8b5-422b-43c6-825e-52ea98516e6c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.417949] env[63372]: DEBUG nova.network.neutron [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 571.419578] env[63372]: DEBUG nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 571.466242] env[63372]: DEBUG nova.virt.hardware [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 571.466242] env[63372]: DEBUG nova.virt.hardware [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 571.466242] env[63372]: DEBUG nova.virt.hardware [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 571.466500] env[63372]: DEBUG nova.virt.hardware [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 571.466500] env[63372]: DEBUG nova.virt.hardware [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 571.466500] env[63372]: DEBUG nova.virt.hardware [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 571.466500] env[63372]: DEBUG nova.virt.hardware [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 571.466718] env[63372]: DEBUG nova.virt.hardware [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 571.466926] env[63372]: DEBUG nova.virt.hardware [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 571.467132] env[63372]: DEBUG nova.virt.hardware [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 571.467335] env[63372]: DEBUG nova.virt.hardware [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 571.469146] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165d854c-0960-40ae-94fc-5a1854c53e9f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.473504] env[63372]: DEBUG nova.network.neutron [req-59c78604-a2a5-46f1-81e2-ae07652cbc87 req-0f39c39d-d091-4b32-8596-f81dbc9b9100 service nova] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.482282] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb0320d3-0ec9-4240-b20b-502dd7c0fe04 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.968991] env[63372]: DEBUG nova.network.neutron [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.976030] env[63372]: DEBUG oslo_concurrency.lockutils [req-59c78604-a2a5-46f1-81e2-ae07652cbc87 req-0f39c39d-d091-4b32-8596-f81dbc9b9100 service nova] Releasing lock "refresh_cache-dc9c4a56-e0af-473b-b47b-2061fe89c6d6" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.976279] env[63372]: DEBUG nova.compute.manager [req-59c78604-a2a5-46f1-81e2-ae07652cbc87 req-0f39c39d-d091-4b32-8596-f81dbc9b9100 service nova] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Received event network-vif-deleted-32a3c00b-7133-4607-b512-85a673630d43 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 572.091581] env[63372]: ERROR nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 92a840c0-9f37-4af2-9b74-f20b406f353d, please check neutron logs for more information. [ 572.091581] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 572.091581] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.091581] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 572.091581] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 572.091581] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 572.091581] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 572.091581] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 572.091581] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.091581] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 572.091581] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.091581] env[63372]: ERROR nova.compute.manager raise self.value [ 572.091581] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 572.091581] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 572.091581] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.091581] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 572.092279] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.092279] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 572.092279] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 92a840c0-9f37-4af2-9b74-f20b406f353d, please check neutron logs for more information. [ 572.092279] env[63372]: ERROR nova.compute.manager [ 572.092279] env[63372]: Traceback (most recent call last): [ 572.092279] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 572.092279] env[63372]: listener.cb(fileno) [ 572.092279] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.092279] env[63372]: result = function(*args, **kwargs) [ 572.092279] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.092279] env[63372]: return func(*args, **kwargs) [ 572.092279] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.092279] env[63372]: raise e [ 572.092279] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.092279] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 572.092279] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 572.092279] env[63372]: created_port_ids = self._update_ports_for_instance( [ 572.092279] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 572.092279] env[63372]: with excutils.save_and_reraise_exception(): [ 572.092279] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.092279] env[63372]: self.force_reraise() [ 572.092279] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.092279] env[63372]: raise self.value [ 572.092279] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 572.092279] env[63372]: updated_port = self._update_port( [ 572.092279] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.092279] env[63372]: _ensure_no_port_binding_failure(port) [ 572.092279] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.092279] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 572.093204] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 92a840c0-9f37-4af2-9b74-f20b406f353d, please check neutron logs for more information. [ 572.093204] env[63372]: Removing descriptor: 14 [ 572.093204] env[63372]: ERROR nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 92a840c0-9f37-4af2-9b74-f20b406f353d, please check neutron logs for more information. [ 572.093204] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Traceback (most recent call last): [ 572.093204] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 572.093204] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] yield resources [ 572.093204] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.093204] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] self.driver.spawn(context, instance, image_meta, [ 572.093204] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 572.093204] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.093204] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.093204] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] vm_ref = self.build_virtual_machine(instance, [ 572.093638] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.093638] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.093638] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.093638] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] for vif in network_info: [ 572.093638] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.093638] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] return self._sync_wrapper(fn, *args, **kwargs) [ 572.093638] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.093638] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] self.wait() [ 572.093638] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.093638] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] self[:] = self._gt.wait() [ 572.093638] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.093638] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] return self._exit_event.wait() [ 572.093638] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.094059] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] result = hub.switch() [ 572.094059] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.094059] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] return self.greenlet.switch() [ 572.094059] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.094059] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] result = function(*args, **kwargs) [ 572.094059] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.094059] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] return func(*args, **kwargs) [ 572.094059] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.094059] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] raise e [ 572.094059] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.094059] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] nwinfo = self.network_api.allocate_for_instance( [ 572.094059] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 572.094059] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] created_port_ids = self._update_ports_for_instance( [ 572.094519] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 572.094519] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] with excutils.save_and_reraise_exception(): [ 572.094519] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.094519] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] self.force_reraise() [ 572.094519] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.094519] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] raise self.value [ 572.094519] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 572.094519] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] updated_port = self._update_port( [ 572.094519] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.094519] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] _ensure_no_port_binding_failure(port) [ 572.094519] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.094519] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] raise exception.PortBindingFailed(port_id=port['id']) [ 572.094914] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] nova.exception.PortBindingFailed: Binding failed for port 92a840c0-9f37-4af2-9b74-f20b406f353d, please check neutron logs for more information. [ 572.094914] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] [ 572.094914] env[63372]: INFO nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Terminating instance [ 572.097017] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Acquiring lock "refresh_cache-392f418f-c8c7-41e3-9bb6-f68e1cab75f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.097017] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Acquired lock "refresh_cache-392f418f-c8c7-41e3-9bb6-f68e1cab75f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.098367] env[63372]: DEBUG nova.network.neutron [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 572.221986] env[63372]: DEBUG nova.network.neutron [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.460313] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bcfed93-ccab-4ce1-b280-f15a6c40819c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.475067] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e73927-ef33-4720-b037-353bae260447 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.518183] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-446d0a9f-680f-4b3f-a6ec-4361772afd7c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.531174] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c9972b-5602-4c22-aeb4-a0a44d30c33e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.543071] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "ac90a156-be00-4f62-a76e-e08914531167" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.543312] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "ac90a156-be00-4f62-a76e-e08914531167" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.556573] env[63372]: DEBUG nova.compute.provider_tree [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.650758] env[63372]: DEBUG nova.network.neutron [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 572.727235] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Releasing lock "refresh_cache-d290c8b5-422b-43c6-825e-52ea98516e6c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.728952] env[63372]: DEBUG nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 572.728952] env[63372]: DEBUG nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 572.728952] env[63372]: DEBUG nova.network.neutron [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 572.754096] env[63372]: DEBUG nova.network.neutron [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.760220] env[63372]: DEBUG nova.network.neutron [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.059598] env[63372]: DEBUG nova.scheduler.client.report [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 573.210260] env[63372]: DEBUG nova.compute.manager [req-459c3607-ce1b-4890-9915-9cfd65ef4ab6 req-d6d4c2ec-4f17-47db-bb40-5052846d808f service nova] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Received event network-changed-92a840c0-9f37-4af2-9b74-f20b406f353d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 573.210545] env[63372]: DEBUG nova.compute.manager [req-459c3607-ce1b-4890-9915-9cfd65ef4ab6 req-d6d4c2ec-4f17-47db-bb40-5052846d808f service nova] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Refreshing instance network info cache due to event network-changed-92a840c0-9f37-4af2-9b74-f20b406f353d. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 573.210623] env[63372]: DEBUG oslo_concurrency.lockutils [req-459c3607-ce1b-4890-9915-9cfd65ef4ab6 req-d6d4c2ec-4f17-47db-bb40-5052846d808f service nova] Acquiring lock "refresh_cache-392f418f-c8c7-41e3-9bb6-f68e1cab75f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.219053] env[63372]: ERROR nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 64490f1e-3e00-4fc9-92fb-a2e454998270, please check neutron logs for more information. [ 573.219053] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 573.219053] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.219053] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 573.219053] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.219053] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 573.219053] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.219053] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 573.219053] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.219053] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 573.219053] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.219053] env[63372]: ERROR nova.compute.manager raise self.value [ 573.219053] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.219053] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 573.219053] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.219053] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 573.219574] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.219574] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 573.219574] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 64490f1e-3e00-4fc9-92fb-a2e454998270, please check neutron logs for more information. [ 573.219574] env[63372]: ERROR nova.compute.manager [ 573.219574] env[63372]: Traceback (most recent call last): [ 573.219574] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 573.219574] env[63372]: listener.cb(fileno) [ 573.219574] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.219574] env[63372]: result = function(*args, **kwargs) [ 573.219574] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.219574] env[63372]: return func(*args, **kwargs) [ 573.219574] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.219574] env[63372]: raise e [ 573.219574] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.219574] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 573.219574] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.219574] env[63372]: created_port_ids = self._update_ports_for_instance( [ 573.219574] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.219574] env[63372]: with excutils.save_and_reraise_exception(): [ 573.219574] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.219574] env[63372]: self.force_reraise() [ 573.219574] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.219574] env[63372]: raise self.value [ 573.219574] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.219574] env[63372]: updated_port = self._update_port( [ 573.219574] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.219574] env[63372]: _ensure_no_port_binding_failure(port) [ 573.219574] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.219574] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 573.220471] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 64490f1e-3e00-4fc9-92fb-a2e454998270, please check neutron logs for more information. [ 573.220471] env[63372]: Removing descriptor: 19 [ 573.220471] env[63372]: ERROR nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 64490f1e-3e00-4fc9-92fb-a2e454998270, please check neutron logs for more information. [ 573.220471] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Traceback (most recent call last): [ 573.220471] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 573.220471] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] yield resources [ 573.220471] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.220471] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] self.driver.spawn(context, instance, image_meta, [ 573.220471] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 573.220471] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.220471] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.220471] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] vm_ref = self.build_virtual_machine(instance, [ 573.220864] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.220864] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.220864] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.220864] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] for vif in network_info: [ 573.220864] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.220864] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] return self._sync_wrapper(fn, *args, **kwargs) [ 573.220864] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.220864] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] self.wait() [ 573.220864] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.220864] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] self[:] = self._gt.wait() [ 573.220864] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.220864] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] return self._exit_event.wait() [ 573.220864] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.221267] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] result = hub.switch() [ 573.221267] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.221267] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] return self.greenlet.switch() [ 573.221267] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.221267] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] result = function(*args, **kwargs) [ 573.221267] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.221267] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] return func(*args, **kwargs) [ 573.221267] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.221267] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] raise e [ 573.221267] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.221267] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] nwinfo = self.network_api.allocate_for_instance( [ 573.221267] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.221267] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] created_port_ids = self._update_ports_for_instance( [ 573.221661] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.221661] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] with excutils.save_and_reraise_exception(): [ 573.221661] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.221661] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] self.force_reraise() [ 573.221661] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.221661] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] raise self.value [ 573.221661] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.221661] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] updated_port = self._update_port( [ 573.221661] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.221661] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] _ensure_no_port_binding_failure(port) [ 573.221661] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.221661] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] raise exception.PortBindingFailed(port_id=port['id']) [ 573.222044] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] nova.exception.PortBindingFailed: Binding failed for port 64490f1e-3e00-4fc9-92fb-a2e454998270, please check neutron logs for more information. [ 573.222044] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] [ 573.222044] env[63372]: INFO nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Terminating instance [ 573.223761] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "refresh_cache-9d52ebd0-f38e-4825-a974-f9a0c0643116" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.223761] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquired lock "refresh_cache-9d52ebd0-f38e-4825-a974-f9a0c0643116" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.223761] env[63372]: DEBUG nova.network.neutron [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 573.256951] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Releasing lock "refresh_cache-392f418f-c8c7-41e3-9bb6-f68e1cab75f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.257463] env[63372]: DEBUG nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 573.257663] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 573.258054] env[63372]: DEBUG oslo_concurrency.lockutils [req-459c3607-ce1b-4890-9915-9cfd65ef4ab6 req-d6d4c2ec-4f17-47db-bb40-5052846d808f service nova] Acquired lock "refresh_cache-392f418f-c8c7-41e3-9bb6-f68e1cab75f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.258489] env[63372]: DEBUG nova.network.neutron [req-459c3607-ce1b-4890-9915-9cfd65ef4ab6 req-d6d4c2ec-4f17-47db-bb40-5052846d808f service nova] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Refreshing network info cache for port 92a840c0-9f37-4af2-9b74-f20b406f353d {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 573.259181] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9af2c43f-ab8c-4c2f-b7d5-5c953f3505a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.268495] env[63372]: DEBUG nova.network.neutron [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.282605] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ce382b-dbca-406a-8851-c61da069545a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.305916] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 392f418f-c8c7-41e3-9bb6-f68e1cab75f3 could not be found. [ 573.305916] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 573.305916] env[63372]: INFO nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Took 0.05 seconds to destroy the instance on the hypervisor. [ 573.306121] env[63372]: DEBUG oslo.service.loopingcall [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 573.306323] env[63372]: DEBUG nova.compute.manager [-] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 573.306323] env[63372]: DEBUG nova.network.neutron [-] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 573.329717] env[63372]: DEBUG nova.network.neutron [-] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.566534] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.155s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.567187] env[63372]: ERROR nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2bdb48e9-8cdf-4e6d-b8ca-12d83d6a2b26, please check neutron logs for more information. [ 573.567187] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Traceback (most recent call last): [ 573.567187] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.567187] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] self.driver.spawn(context, instance, image_meta, [ 573.567187] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 573.567187] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.567187] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.567187] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] vm_ref = self.build_virtual_machine(instance, [ 573.567187] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.567187] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.567187] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.567720] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] for vif in network_info: [ 573.567720] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.567720] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] return self._sync_wrapper(fn, *args, **kwargs) [ 573.567720] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.567720] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] self.wait() [ 573.567720] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.567720] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] self[:] = self._gt.wait() [ 573.567720] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.567720] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] return self._exit_event.wait() [ 573.567720] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.567720] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] result = hub.switch() [ 573.567720] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.567720] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] return self.greenlet.switch() [ 573.568233] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.568233] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] result = function(*args, **kwargs) [ 573.568233] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.568233] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] return func(*args, **kwargs) [ 573.568233] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.568233] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] raise e [ 573.568233] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.568233] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] nwinfo = self.network_api.allocate_for_instance( [ 573.568233] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.568233] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] created_port_ids = self._update_ports_for_instance( [ 573.568233] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.568233] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] with excutils.save_and_reraise_exception(): [ 573.568233] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.568692] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] self.force_reraise() [ 573.568692] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.568692] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] raise self.value [ 573.568692] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.568692] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] updated_port = self._update_port( [ 573.568692] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.568692] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] _ensure_no_port_binding_failure(port) [ 573.568692] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.568692] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] raise exception.PortBindingFailed(port_id=port['id']) [ 573.568692] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] nova.exception.PortBindingFailed: Binding failed for port 2bdb48e9-8cdf-4e6d-b8ca-12d83d6a2b26, please check neutron logs for more information. [ 573.568692] env[63372]: ERROR nova.compute.manager [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] [ 573.569129] env[63372]: DEBUG nova.compute.utils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Binding failed for port 2bdb48e9-8cdf-4e6d-b8ca-12d83d6a2b26, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 573.569299] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.011s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.572165] env[63372]: DEBUG nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Build of instance 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44 was re-scheduled: Binding failed for port 2bdb48e9-8cdf-4e6d-b8ca-12d83d6a2b26, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 573.572639] env[63372]: DEBUG nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 573.572880] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Acquiring lock "refresh_cache-292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.573043] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Acquired lock "refresh_cache-292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.573207] env[63372]: DEBUG nova.network.neutron [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 573.744199] env[63372]: DEBUG nova.network.neutron [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.771923] env[63372]: INFO nova.compute.manager [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] [instance: d290c8b5-422b-43c6-825e-52ea98516e6c] Took 1.04 seconds to deallocate network for instance. [ 573.794169] env[63372]: DEBUG nova.network.neutron [req-459c3607-ce1b-4890-9915-9cfd65ef4ab6 req-d6d4c2ec-4f17-47db-bb40-5052846d808f service nova] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.831635] env[63372]: DEBUG nova.network.neutron [-] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.877563] env[63372]: DEBUG nova.network.neutron [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.923820] env[63372]: DEBUG nova.network.neutron [req-459c3607-ce1b-4890-9915-9cfd65ef4ab6 req-d6d4c2ec-4f17-47db-bb40-5052846d808f service nova] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.105577] env[63372]: DEBUG nova.network.neutron [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.240820] env[63372]: DEBUG nova.network.neutron [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.260595] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "5819c38e-2cf9-4d16-b28a-5f23d35c3d44" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.260904] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "5819c38e-2cf9-4d16-b28a-5f23d35c3d44" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.336893] env[63372]: INFO nova.compute.manager [-] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Took 1.03 seconds to deallocate network for instance. [ 574.339690] env[63372]: DEBUG nova.compute.claims [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 574.339856] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.381171] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Releasing lock "refresh_cache-9d52ebd0-f38e-4825-a974-f9a0c0643116" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.381612] env[63372]: DEBUG nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 574.381811] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 574.382106] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a5a90dae-b5fd-44b7-be0f-05c8c03a82aa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.394418] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b96859a5-cee5-4740-a81e-a0c6230e393b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.421610] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9d52ebd0-f38e-4825-a974-f9a0c0643116 could not be found. [ 574.421862] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 574.422077] env[63372]: INFO nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Took 0.04 seconds to destroy the instance on the hypervisor. [ 574.422329] env[63372]: DEBUG oslo.service.loopingcall [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.425555] env[63372]: DEBUG nova.compute.manager [-] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.425655] env[63372]: DEBUG nova.network.neutron [-] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 574.429110] env[63372]: DEBUG oslo_concurrency.lockutils [req-459c3607-ce1b-4890-9915-9cfd65ef4ab6 req-d6d4c2ec-4f17-47db-bb40-5052846d808f service nova] Releasing lock "refresh_cache-392f418f-c8c7-41e3-9bb6-f68e1cab75f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.429110] env[63372]: DEBUG nova.compute.manager [req-459c3607-ce1b-4890-9915-9cfd65ef4ab6 req-d6d4c2ec-4f17-47db-bb40-5052846d808f service nova] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Received event network-vif-deleted-92a840c0-9f37-4af2-9b74-f20b406f353d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 574.451301] env[63372]: DEBUG nova.network.neutron [-] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.543590] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf3f87e-bd39-4a55-9a48-0ef960e8bda1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.552875] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10c3e29-a6d4-4847-ba03-eff1973671ba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.589639] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43acf6cf-2464-489d-81b7-eec14ca18fab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.602389] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f078c727-3cae-415c-be7f-8983b17ce024 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.619291] env[63372]: DEBUG nova.compute.provider_tree [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.746660] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Releasing lock "refresh_cache-292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.746894] env[63372]: DEBUG nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 574.747094] env[63372]: DEBUG nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.747268] env[63372]: DEBUG nova.network.neutron [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 574.774076] env[63372]: DEBUG nova.network.neutron [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.812042] env[63372]: INFO nova.scheduler.client.report [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Deleted allocations for instance d290c8b5-422b-43c6-825e-52ea98516e6c [ 574.955686] env[63372]: DEBUG nova.network.neutron [-] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.128027] env[63372]: DEBUG nova.scheduler.client.report [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 575.280707] env[63372]: DEBUG nova.network.neutron [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.323376] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6c4ba2a-3b72-48e1-8bec-7996f26420cc tempest-AttachInterfacesV270Test-841796072 tempest-AttachInterfacesV270Test-841796072-project-member] Lock "d290c8b5-422b-43c6-825e-52ea98516e6c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.972s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.461158] env[63372]: INFO nova.compute.manager [-] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Took 1.04 seconds to deallocate network for instance. [ 575.465265] env[63372]: DEBUG nova.compute.claims [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 575.465419] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.631633] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.062s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.632217] env[63372]: ERROR nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1e3d9db7-b8f0-4221-9a0f-8538208cb01b, please check neutron logs for more information. [ 575.632217] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Traceback (most recent call last): [ 575.632217] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.632217] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] self.driver.spawn(context, instance, image_meta, [ 575.632217] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 575.632217] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.632217] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.632217] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] vm_ref = self.build_virtual_machine(instance, [ 575.632217] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.632217] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.632217] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.632719] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] for vif in network_info: [ 575.632719] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.632719] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] return self._sync_wrapper(fn, *args, **kwargs) [ 575.632719] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.632719] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] self.wait() [ 575.632719] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.632719] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] self[:] = self._gt.wait() [ 575.632719] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.632719] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] return self._exit_event.wait() [ 575.632719] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.632719] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] result = hub.switch() [ 575.632719] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.632719] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] return self.greenlet.switch() [ 575.633283] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.633283] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] result = function(*args, **kwargs) [ 575.633283] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.633283] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] return func(*args, **kwargs) [ 575.633283] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.633283] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] raise e [ 575.633283] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.633283] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] nwinfo = self.network_api.allocate_for_instance( [ 575.633283] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.633283] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] created_port_ids = self._update_ports_for_instance( [ 575.633283] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.633283] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] with excutils.save_and_reraise_exception(): [ 575.633283] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.633748] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] self.force_reraise() [ 575.633748] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.633748] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] raise self.value [ 575.633748] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.633748] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] updated_port = self._update_port( [ 575.633748] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.633748] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] _ensure_no_port_binding_failure(port) [ 575.633748] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.633748] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] raise exception.PortBindingFailed(port_id=port['id']) [ 575.633748] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] nova.exception.PortBindingFailed: Binding failed for port 1e3d9db7-b8f0-4221-9a0f-8538208cb01b, please check neutron logs for more information. [ 575.633748] env[63372]: ERROR nova.compute.manager [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] [ 575.634171] env[63372]: DEBUG nova.compute.utils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Binding failed for port 1e3d9db7-b8f0-4221-9a0f-8538208cb01b, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 575.635321] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.472s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.639550] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Build of instance ee2f5088-616b-4237-93c9-fac5e0d6aba2 was re-scheduled: Binding failed for port 1e3d9db7-b8f0-4221-9a0f-8538208cb01b, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 575.641627] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 575.641627] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "refresh_cache-ee2f5088-616b-4237-93c9-fac5e0d6aba2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.641627] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquired lock "refresh_cache-ee2f5088-616b-4237-93c9-fac5e0d6aba2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.641627] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 575.741412] env[63372]: DEBUG nova.compute.manager [req-62aae2a7-ec2a-4079-b6a8-b353896d03e0 req-5ae0874b-10e3-4e9d-8c28-9897f415c1b7 service nova] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Received event network-changed-64490f1e-3e00-4fc9-92fb-a2e454998270 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 575.741482] env[63372]: DEBUG nova.compute.manager [req-62aae2a7-ec2a-4079-b6a8-b353896d03e0 req-5ae0874b-10e3-4e9d-8c28-9897f415c1b7 service nova] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Refreshing instance network info cache due to event network-changed-64490f1e-3e00-4fc9-92fb-a2e454998270. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 575.741825] env[63372]: DEBUG oslo_concurrency.lockutils [req-62aae2a7-ec2a-4079-b6a8-b353896d03e0 req-5ae0874b-10e3-4e9d-8c28-9897f415c1b7 service nova] Acquiring lock "refresh_cache-9d52ebd0-f38e-4825-a974-f9a0c0643116" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.741909] env[63372]: DEBUG oslo_concurrency.lockutils [req-62aae2a7-ec2a-4079-b6a8-b353896d03e0 req-5ae0874b-10e3-4e9d-8c28-9897f415c1b7 service nova] Acquired lock "refresh_cache-9d52ebd0-f38e-4825-a974-f9a0c0643116" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.742073] env[63372]: DEBUG nova.network.neutron [req-62aae2a7-ec2a-4079-b6a8-b353896d03e0 req-5ae0874b-10e3-4e9d-8c28-9897f415c1b7 service nova] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Refreshing network info cache for port 64490f1e-3e00-4fc9-92fb-a2e454998270 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 575.783663] env[63372]: INFO nova.compute.manager [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] [instance: 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44] Took 1.04 seconds to deallocate network for instance. [ 575.826029] env[63372]: DEBUG nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 576.172150] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.266682] env[63372]: DEBUG nova.network.neutron [req-62aae2a7-ec2a-4079-b6a8-b353896d03e0 req-5ae0874b-10e3-4e9d-8c28-9897f415c1b7 service nova] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.316127] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.353724] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.389577] env[63372]: DEBUG nova.network.neutron [req-62aae2a7-ec2a-4079-b6a8-b353896d03e0 req-5ae0874b-10e3-4e9d-8c28-9897f415c1b7 service nova] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.622541] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aac4bac-8269-44dd-8570-5d39c29150a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.631885] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe271ef3-5bfd-4d90-a1e3-5198ed8f116b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.667606] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c1b1b2-b1e2-43ac-aa6a-8ecb462a17cd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.675486] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0928b126-5105-436d-a13a-8cc7acf9fa6e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.692024] env[63372]: DEBUG nova.compute.provider_tree [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.819612] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Releasing lock "refresh_cache-ee2f5088-616b-4237-93c9-fac5e0d6aba2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.819693] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 576.820163] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 576.820163] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 576.823131] env[63372]: INFO nova.scheduler.client.report [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Deleted allocations for instance 292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44 [ 576.852076] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.893350] env[63372]: DEBUG oslo_concurrency.lockutils [req-62aae2a7-ec2a-4079-b6a8-b353896d03e0 req-5ae0874b-10e3-4e9d-8c28-9897f415c1b7 service nova] Releasing lock "refresh_cache-9d52ebd0-f38e-4825-a974-f9a0c0643116" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.894157] env[63372]: DEBUG nova.compute.manager [req-62aae2a7-ec2a-4079-b6a8-b353896d03e0 req-5ae0874b-10e3-4e9d-8c28-9897f415c1b7 service nova] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Received event network-vif-deleted-64490f1e-3e00-4fc9-92fb-a2e454998270 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 577.197297] env[63372]: DEBUG nova.scheduler.client.report [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.332451] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46ba0b5d-9cb4-49cc-ba5d-cfc942f6f807 tempest-ServerActionsTestOtherA-1499166749 tempest-ServerActionsTestOtherA-1499166749-project-member] Lock "292d8d26-cae8-4e18-ba4f-0dc8d5ea2f44" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.339s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.354775] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.703553] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.069s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.704191] env[63372]: ERROR nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 66605f0f-5cf3-4815-9309-0d3bcd4144d8, please check neutron logs for more information. [ 577.704191] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Traceback (most recent call last): [ 577.704191] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.704191] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] self.driver.spawn(context, instance, image_meta, [ 577.704191] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 577.704191] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.704191] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.704191] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] vm_ref = self.build_virtual_machine(instance, [ 577.704191] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.704191] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.704191] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.704755] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] for vif in network_info: [ 577.704755] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.704755] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] return self._sync_wrapper(fn, *args, **kwargs) [ 577.704755] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.704755] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] self.wait() [ 577.704755] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.704755] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] self[:] = self._gt.wait() [ 577.704755] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.704755] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] return self._exit_event.wait() [ 577.704755] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.704755] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] result = hub.switch() [ 577.704755] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.704755] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] return self.greenlet.switch() [ 577.705171] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.705171] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] result = function(*args, **kwargs) [ 577.705171] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.705171] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] return func(*args, **kwargs) [ 577.705171] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.705171] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] raise e [ 577.705171] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.705171] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] nwinfo = self.network_api.allocate_for_instance( [ 577.705171] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.705171] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] created_port_ids = self._update_ports_for_instance( [ 577.705171] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.705171] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] with excutils.save_and_reraise_exception(): [ 577.705171] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.705618] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] self.force_reraise() [ 577.705618] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.705618] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] raise self.value [ 577.705618] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.705618] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] updated_port = self._update_port( [ 577.705618] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.705618] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] _ensure_no_port_binding_failure(port) [ 577.705618] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.705618] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] raise exception.PortBindingFailed(port_id=port['id']) [ 577.705618] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] nova.exception.PortBindingFailed: Binding failed for port 66605f0f-5cf3-4815-9309-0d3bcd4144d8, please check neutron logs for more information. [ 577.705618] env[63372]: ERROR nova.compute.manager [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] [ 577.705927] env[63372]: DEBUG nova.compute.utils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Binding failed for port 66605f0f-5cf3-4815-9309-0d3bcd4144d8, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 577.707029] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.415s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.708560] env[63372]: INFO nova.compute.claims [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 577.715640] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Build of instance 39e22917-3ea8-457c-9d6e-679a848c3fd3 was re-scheduled: Binding failed for port 66605f0f-5cf3-4815-9309-0d3bcd4144d8, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 577.716169] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 577.716402] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "refresh_cache-39e22917-3ea8-457c-9d6e-679a848c3fd3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.716546] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquired lock "refresh_cache-39e22917-3ea8-457c-9d6e-679a848c3fd3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.716705] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 577.835214] env[63372]: DEBUG nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 577.860051] env[63372]: INFO nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: ee2f5088-616b-4237-93c9-fac5e0d6aba2] Took 1.04 seconds to deallocate network for instance. [ 578.246663] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.363960] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.375715] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.572712] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "c4718797-aa86-4ec0-94d3-6480bd6aa898" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.572993] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "c4718797-aa86-4ec0-94d3-6480bd6aa898" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.879698] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Releasing lock "refresh_cache-39e22917-3ea8-457c-9d6e-679a848c3fd3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.879698] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 578.879698] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 578.879698] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 578.894441] env[63372]: INFO nova.scheduler.client.report [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Deleted allocations for instance ee2f5088-616b-4237-93c9-fac5e0d6aba2 [ 578.910125] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 579.178738] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c731549c-f216-47db-9f3c-8c8007ab97e2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.188322] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e8ef75a-09bf-4f83-bc40-d4d5e6410cd1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.224655] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d74bb5-26b1-42fc-86a1-2e9f215509b0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.232965] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce757087-9de4-46a1-86d9-09e48063bb93 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.248418] env[63372]: DEBUG nova.compute.provider_tree [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.412723] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "ee2f5088-616b-4237-93c9-fac5e0d6aba2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.542s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.413322] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.752028] env[63372]: DEBUG nova.scheduler.client.report [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.918970] env[63372]: INFO nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 39e22917-3ea8-457c-9d6e-679a848c3fd3] Took 1.04 seconds to deallocate network for instance. [ 579.927280] env[63372]: DEBUG nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 580.259661] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.552s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.260221] env[63372]: DEBUG nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 580.266068] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.218s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.472033] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.772914] env[63372]: DEBUG nova.compute.utils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 580.777492] env[63372]: DEBUG nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 580.778683] env[63372]: DEBUG nova.network.neutron [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 580.950103] env[63372]: DEBUG nova.policy [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1625f9903b84460a8ee0318dda7adaad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b0dabe1be5324f31b3d7ee3d40352950', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 580.991825] env[63372]: INFO nova.scheduler.client.report [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Deleted allocations for instance 39e22917-3ea8-457c-9d6e-679a848c3fd3 [ 581.279516] env[63372]: DEBUG nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 581.314856] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2783b0a5-fa52-403f-a332-dcbc55d5879a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.322595] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abab0e2a-7c37-4d22-b2aa-83edc2c9a369 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.357022] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81560374-d14d-4f81-b4b9-e3887316b008 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.363992] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eda6174-8b48-4977-92b6-0ef74a239e27 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.380299] env[63372]: DEBUG nova.compute.provider_tree [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.508669] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "39e22917-3ea8-457c-9d6e-679a848c3fd3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.607s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.590265] env[63372]: DEBUG nova.network.neutron [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Successfully created port: 7c7784d9-31f5-4c0a-b5d8-93d147d846c0 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 581.883289] env[63372]: DEBUG nova.scheduler.client.report [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.013149] env[63372]: DEBUG nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 582.291439] env[63372]: DEBUG nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 582.323879] env[63372]: DEBUG nova.virt.hardware [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 582.324158] env[63372]: DEBUG nova.virt.hardware [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 582.324315] env[63372]: DEBUG nova.virt.hardware [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 582.324494] env[63372]: DEBUG nova.virt.hardware [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 582.324689] env[63372]: DEBUG nova.virt.hardware [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 582.324775] env[63372]: DEBUG nova.virt.hardware [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 582.325110] env[63372]: DEBUG nova.virt.hardware [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 582.325346] env[63372]: DEBUG nova.virt.hardware [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 582.325530] env[63372]: DEBUG nova.virt.hardware [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 582.325695] env[63372]: DEBUG nova.virt.hardware [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 582.325873] env[63372]: DEBUG nova.virt.hardware [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 582.327320] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-badd946e-7fb6-4a79-b4cf-77ce14338368 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.335959] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02121002-49f0-47b2-9978-33c011bfde07 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.392226] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.128s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.392888] env[63372]: ERROR nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9eb8da80-70b2-45fe-92f5-022d9b9a217c, please check neutron logs for more information. [ 582.392888] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Traceback (most recent call last): [ 582.392888] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.392888] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] self.driver.spawn(context, instance, image_meta, [ 582.392888] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 582.392888] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.392888] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.392888] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] vm_ref = self.build_virtual_machine(instance, [ 582.392888] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.392888] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.392888] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.393273] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] for vif in network_info: [ 582.393273] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.393273] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] return self._sync_wrapper(fn, *args, **kwargs) [ 582.393273] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.393273] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] self.wait() [ 582.393273] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.393273] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] self[:] = self._gt.wait() [ 582.393273] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.393273] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] return self._exit_event.wait() [ 582.393273] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.393273] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] result = hub.switch() [ 582.393273] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.393273] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] return self.greenlet.switch() [ 582.393633] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.393633] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] result = function(*args, **kwargs) [ 582.393633] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.393633] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] return func(*args, **kwargs) [ 582.393633] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.393633] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] raise e [ 582.393633] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.393633] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] nwinfo = self.network_api.allocate_for_instance( [ 582.393633] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.393633] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] created_port_ids = self._update_ports_for_instance( [ 582.393633] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.393633] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] with excutils.save_and_reraise_exception(): [ 582.393633] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.393978] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] self.force_reraise() [ 582.393978] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.393978] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] raise self.value [ 582.393978] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.393978] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] updated_port = self._update_port( [ 582.393978] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.393978] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] _ensure_no_port_binding_failure(port) [ 582.393978] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.393978] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] raise exception.PortBindingFailed(port_id=port['id']) [ 582.393978] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] nova.exception.PortBindingFailed: Binding failed for port 9eb8da80-70b2-45fe-92f5-022d9b9a217c, please check neutron logs for more information. [ 582.393978] env[63372]: ERROR nova.compute.manager [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] [ 582.394328] env[63372]: DEBUG nova.compute.utils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Binding failed for port 9eb8da80-70b2-45fe-92f5-022d9b9a217c, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 582.394829] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.758s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.399962] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Build of instance 7803b83f-2944-4e3b-87b1-0d3cea3455ad was re-scheduled: Binding failed for port 9eb8da80-70b2-45fe-92f5-022d9b9a217c, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 582.400787] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 582.401103] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquiring lock "refresh_cache-7803b83f-2944-4e3b-87b1-0d3cea3455ad" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.401286] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Acquired lock "refresh_cache-7803b83f-2944-4e3b-87b1-0d3cea3455ad" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.401426] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 582.543195] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.937211] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.182936] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.343258] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a244328-6e23-4b31-905c-902477a39670 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.352559] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53fc1dfe-0416-4a6f-aaa6-745987ff68a3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.387831] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7395b7df-1dfd-4d87-bbfd-7b16d6513d3a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.395802] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854b761d-b255-40db-b191-ce54cc86f1e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.413345] env[63372]: DEBUG nova.compute.provider_tree [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.687046] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Releasing lock "refresh_cache-7803b83f-2944-4e3b-87b1-0d3cea3455ad" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.687343] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 583.687490] env[63372]: DEBUG nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.687644] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 583.724611] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.916156] env[63372]: DEBUG nova.scheduler.client.report [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.058310] env[63372]: ERROR nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7c7784d9-31f5-4c0a-b5d8-93d147d846c0, please check neutron logs for more information. [ 584.058310] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 584.058310] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.058310] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 584.058310] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.058310] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 584.058310] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.058310] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 584.058310] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.058310] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 584.058310] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.058310] env[63372]: ERROR nova.compute.manager raise self.value [ 584.058310] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.058310] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 584.058310] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.058310] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 584.058880] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.058880] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 584.058880] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7c7784d9-31f5-4c0a-b5d8-93d147d846c0, please check neutron logs for more information. [ 584.058880] env[63372]: ERROR nova.compute.manager [ 584.058880] env[63372]: Traceback (most recent call last): [ 584.058880] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 584.058880] env[63372]: listener.cb(fileno) [ 584.058880] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.058880] env[63372]: result = function(*args, **kwargs) [ 584.058880] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.058880] env[63372]: return func(*args, **kwargs) [ 584.058880] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.058880] env[63372]: raise e [ 584.058880] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.058880] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 584.058880] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.058880] env[63372]: created_port_ids = self._update_ports_for_instance( [ 584.058880] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.058880] env[63372]: with excutils.save_and_reraise_exception(): [ 584.058880] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.058880] env[63372]: self.force_reraise() [ 584.058880] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.058880] env[63372]: raise self.value [ 584.058880] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.058880] env[63372]: updated_port = self._update_port( [ 584.058880] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.058880] env[63372]: _ensure_no_port_binding_failure(port) [ 584.058880] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.058880] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 584.059706] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 7c7784d9-31f5-4c0a-b5d8-93d147d846c0, please check neutron logs for more information. [ 584.059706] env[63372]: Removing descriptor: 19 [ 584.059706] env[63372]: ERROR nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7c7784d9-31f5-4c0a-b5d8-93d147d846c0, please check neutron logs for more information. [ 584.059706] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Traceback (most recent call last): [ 584.059706] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 584.059706] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] yield resources [ 584.059706] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 584.059706] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] self.driver.spawn(context, instance, image_meta, [ 584.059706] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 584.059706] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.059706] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.059706] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] vm_ref = self.build_virtual_machine(instance, [ 584.060270] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.060270] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.060270] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.060270] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] for vif in network_info: [ 584.060270] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 584.060270] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] return self._sync_wrapper(fn, *args, **kwargs) [ 584.060270] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 584.060270] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] self.wait() [ 584.060270] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 584.060270] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] self[:] = self._gt.wait() [ 584.060270] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.060270] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] return self._exit_event.wait() [ 584.060270] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.060627] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] result = hub.switch() [ 584.060627] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.060627] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] return self.greenlet.switch() [ 584.060627] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.060627] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] result = function(*args, **kwargs) [ 584.060627] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.060627] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] return func(*args, **kwargs) [ 584.060627] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.060627] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] raise e [ 584.060627] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.060627] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] nwinfo = self.network_api.allocate_for_instance( [ 584.060627] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.060627] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] created_port_ids = self._update_ports_for_instance( [ 584.060990] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.060990] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] with excutils.save_and_reraise_exception(): [ 584.060990] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.060990] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] self.force_reraise() [ 584.060990] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.060990] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] raise self.value [ 584.060990] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.060990] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] updated_port = self._update_port( [ 584.060990] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.060990] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] _ensure_no_port_binding_failure(port) [ 584.060990] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.060990] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] raise exception.PortBindingFailed(port_id=port['id']) [ 584.061368] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] nova.exception.PortBindingFailed: Binding failed for port 7c7784d9-31f5-4c0a-b5d8-93d147d846c0, please check neutron logs for more information. [ 584.061368] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] [ 584.061368] env[63372]: INFO nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Terminating instance [ 584.062882] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Acquiring lock "refresh_cache-45d6d393-ca23-4e22-bcbf-899ae5037c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.063066] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Acquired lock "refresh_cache-45d6d393-ca23-4e22-bcbf-899ae5037c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.063233] env[63372]: DEBUG nova.network.neutron [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 584.232671] env[63372]: DEBUG nova.network.neutron [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.354245] env[63372]: DEBUG nova.compute.manager [req-6ffa4bc0-47b4-48e0-8ad9-723a15bc552f req-ce6b0853-ca76-4aa8-aaf4-7167f4adf61c service nova] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Received event network-changed-7c7784d9-31f5-4c0a-b5d8-93d147d846c0 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 584.354245] env[63372]: DEBUG nova.compute.manager [req-6ffa4bc0-47b4-48e0-8ad9-723a15bc552f req-ce6b0853-ca76-4aa8-aaf4-7167f4adf61c service nova] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Refreshing instance network info cache due to event network-changed-7c7784d9-31f5-4c0a-b5d8-93d147d846c0. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 584.354245] env[63372]: DEBUG oslo_concurrency.lockutils [req-6ffa4bc0-47b4-48e0-8ad9-723a15bc552f req-ce6b0853-ca76-4aa8-aaf4-7167f4adf61c service nova] Acquiring lock "refresh_cache-45d6d393-ca23-4e22-bcbf-899ae5037c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.423093] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.026s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.423093] env[63372]: ERROR nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 88693cd4-b712-426d-8db2-67c49dc783d5, please check neutron logs for more information. [ 584.423093] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Traceback (most recent call last): [ 584.423093] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 584.423093] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] self.driver.spawn(context, instance, image_meta, [ 584.423093] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 584.423093] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.423093] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.423093] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] vm_ref = self.build_virtual_machine(instance, [ 584.423443] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.423443] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.423443] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.423443] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] for vif in network_info: [ 584.423443] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 584.423443] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] return self._sync_wrapper(fn, *args, **kwargs) [ 584.423443] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 584.423443] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] self.wait() [ 584.423443] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 584.423443] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] self[:] = self._gt.wait() [ 584.423443] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.423443] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] return self._exit_event.wait() [ 584.423443] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.423864] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] result = hub.switch() [ 584.423864] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.423864] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] return self.greenlet.switch() [ 584.423864] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.423864] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] result = function(*args, **kwargs) [ 584.423864] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.423864] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] return func(*args, **kwargs) [ 584.423864] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.423864] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] raise e [ 584.423864] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.423864] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] nwinfo = self.network_api.allocate_for_instance( [ 584.423864] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.423864] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] created_port_ids = self._update_ports_for_instance( [ 584.424301] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.424301] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] with excutils.save_and_reraise_exception(): [ 584.424301] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.424301] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] self.force_reraise() [ 584.424301] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.424301] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] raise self.value [ 584.424301] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.424301] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] updated_port = self._update_port( [ 584.424301] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.424301] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] _ensure_no_port_binding_failure(port) [ 584.424301] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.424301] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] raise exception.PortBindingFailed(port_id=port['id']) [ 584.425075] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] nova.exception.PortBindingFailed: Binding failed for port 88693cd4-b712-426d-8db2-67c49dc783d5, please check neutron logs for more information. [ 584.425075] env[63372]: ERROR nova.compute.manager [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] [ 584.425075] env[63372]: DEBUG nova.compute.utils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Binding failed for port 88693cd4-b712-426d-8db2-67c49dc783d5, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 584.429239] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.346s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.432635] env[63372]: DEBUG nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Build of instance 9ec742c1-894f-4baa-a451-01ebb4d2816a was re-scheduled: Binding failed for port 88693cd4-b712-426d-8db2-67c49dc783d5, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 584.432635] env[63372]: DEBUG nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 584.432635] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Acquiring lock "refresh_cache-9ec742c1-894f-4baa-a451-01ebb4d2816a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.432635] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Acquired lock "refresh_cache-9ec742c1-894f-4baa-a451-01ebb4d2816a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.433100] env[63372]: DEBUG nova.network.neutron [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 584.647728] env[63372]: DEBUG nova.network.neutron [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 584.734195] env[63372]: INFO nova.compute.manager [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] [instance: 7803b83f-2944-4e3b-87b1-0d3cea3455ad] Took 1.05 seconds to deallocate network for instance. [ 584.749319] env[63372]: DEBUG nova.network.neutron [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.963028] env[63372]: DEBUG nova.network.neutron [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.204819] env[63372]: DEBUG nova.network.neutron [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.254760] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Releasing lock "refresh_cache-45d6d393-ca23-4e22-bcbf-899ae5037c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.255257] env[63372]: DEBUG nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 585.255438] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 585.265950] env[63372]: DEBUG oslo_concurrency.lockutils [req-6ffa4bc0-47b4-48e0-8ad9-723a15bc552f req-ce6b0853-ca76-4aa8-aaf4-7167f4adf61c service nova] Acquired lock "refresh_cache-45d6d393-ca23-4e22-bcbf-899ae5037c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.266188] env[63372]: DEBUG nova.network.neutron [req-6ffa4bc0-47b4-48e0-8ad9-723a15bc552f req-ce6b0853-ca76-4aa8-aaf4-7167f4adf61c service nova] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Refreshing network info cache for port 7c7784d9-31f5-4c0a-b5d8-93d147d846c0 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 585.267303] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-55685c0a-a8eb-42fe-a660-c9855e9ef169 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.287371] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df868f6-e39a-4476-a2ba-c0fca68fdfb3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.324817] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 45d6d393-ca23-4e22-bcbf-899ae5037c4f could not be found. [ 585.324817] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 585.327847] env[63372]: INFO nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Took 0.07 seconds to destroy the instance on the hypervisor. [ 585.327847] env[63372]: DEBUG oslo.service.loopingcall [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 585.327847] env[63372]: DEBUG nova.compute.manager [-] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.327847] env[63372]: DEBUG nova.network.neutron [-] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 585.416666] env[63372]: DEBUG nova.network.neutron [-] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.445285] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6a6dfb-8f5f-48e5-9e04-4cbd2857b726 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.457826] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58127cfc-8f95-4ce3-b8bc-ff87657428e9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.496642] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e7122c2-c10a-442d-a800-e5f357f6155c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.504347] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5468a69-5156-4600-9844-7c7dcb42109c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.520486] env[63372]: DEBUG nova.compute.provider_tree [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.717270] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Releasing lock "refresh_cache-9ec742c1-894f-4baa-a451-01ebb4d2816a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.717270] env[63372]: DEBUG nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 585.717270] env[63372]: DEBUG nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.717270] env[63372]: DEBUG nova.network.neutron [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 585.776124] env[63372]: DEBUG nova.network.neutron [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.799927] env[63372]: DEBUG nova.network.neutron [req-6ffa4bc0-47b4-48e0-8ad9-723a15bc552f req-ce6b0853-ca76-4aa8-aaf4-7167f4adf61c service nova] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.893612] env[63372]: DEBUG nova.network.neutron [req-6ffa4bc0-47b4-48e0-8ad9-723a15bc552f req-ce6b0853-ca76-4aa8-aaf4-7167f4adf61c service nova] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.919102] env[63372]: DEBUG nova.network.neutron [-] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.977460] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.977460] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.028017] env[63372]: DEBUG nova.scheduler.client.report [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.281274] env[63372]: DEBUG nova.network.neutron [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.398188] env[63372]: DEBUG oslo_concurrency.lockutils [req-6ffa4bc0-47b4-48e0-8ad9-723a15bc552f req-ce6b0853-ca76-4aa8-aaf4-7167f4adf61c service nova] Releasing lock "refresh_cache-45d6d393-ca23-4e22-bcbf-899ae5037c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.423056] env[63372]: INFO nova.compute.manager [-] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Took 1.10 seconds to deallocate network for instance. [ 586.425352] env[63372]: DEBUG nova.compute.claims [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 586.425352] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.535167] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.106s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.535167] env[63372]: ERROR nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 32a3c00b-7133-4607-b512-85a673630d43, please check neutron logs for more information. [ 586.535167] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Traceback (most recent call last): [ 586.535167] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.535167] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] self.driver.spawn(context, instance, image_meta, [ 586.535167] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 586.535167] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.535167] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.535167] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] vm_ref = self.build_virtual_machine(instance, [ 586.535574] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.535574] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.535574] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.535574] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] for vif in network_info: [ 586.535574] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.535574] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] return self._sync_wrapper(fn, *args, **kwargs) [ 586.535574] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.535574] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] self.wait() [ 586.535574] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.535574] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] self[:] = self._gt.wait() [ 586.535574] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.535574] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] return self._exit_event.wait() [ 586.535574] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.535998] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] result = hub.switch() [ 586.535998] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.535998] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] return self.greenlet.switch() [ 586.535998] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.535998] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] result = function(*args, **kwargs) [ 586.535998] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.535998] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] return func(*args, **kwargs) [ 586.535998] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.535998] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] raise e [ 586.535998] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.535998] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] nwinfo = self.network_api.allocate_for_instance( [ 586.535998] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 586.535998] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] created_port_ids = self._update_ports_for_instance( [ 586.536435] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 586.536435] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] with excutils.save_and_reraise_exception(): [ 586.536435] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.536435] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] self.force_reraise() [ 586.536435] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.536435] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] raise self.value [ 586.536435] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 586.536435] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] updated_port = self._update_port( [ 586.536435] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.536435] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] _ensure_no_port_binding_failure(port) [ 586.536435] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.536435] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] raise exception.PortBindingFailed(port_id=port['id']) [ 586.536787] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] nova.exception.PortBindingFailed: Binding failed for port 32a3c00b-7133-4607-b512-85a673630d43, please check neutron logs for more information. [ 586.536787] env[63372]: ERROR nova.compute.manager [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] [ 586.536787] env[63372]: DEBUG nova.compute.utils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Binding failed for port 32a3c00b-7133-4607-b512-85a673630d43, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 586.538264] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.195s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.547406] env[63372]: DEBUG nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Build of instance dc9c4a56-e0af-473b-b47b-2061fe89c6d6 was re-scheduled: Binding failed for port 32a3c00b-7133-4607-b512-85a673630d43, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 586.547406] env[63372]: DEBUG nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 586.547406] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Acquiring lock "refresh_cache-dc9c4a56-e0af-473b-b47b-2061fe89c6d6" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.547406] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Acquired lock "refresh_cache-dc9c4a56-e0af-473b-b47b-2061fe89c6d6" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.547601] env[63372]: DEBUG nova.network.neutron [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 586.792418] env[63372]: INFO nova.compute.manager [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: 9ec742c1-894f-4baa-a451-01ebb4d2816a] Took 1.08 seconds to deallocate network for instance. [ 586.805313] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bbea4c6e-0c91-4ade-b817-0558ed9d54fe tempest-ListServersNegativeTestJSON-1595954517 tempest-ListServersNegativeTestJSON-1595954517-project-member] Lock "7803b83f-2944-4e3b-87b1-0d3cea3455ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.864s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.805313] env[63372]: Traceback (most recent call last): [ 586.805313] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.805313] env[63372]: self.driver.spawn(context, instance, image_meta, [ 586.805313] env[63372]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 586.805313] env[63372]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.805313] env[63372]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.805313] env[63372]: vm_ref = self.build_virtual_machine(instance, [ 586.805313] env[63372]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.805313] env[63372]: vif_infos = vmwarevif.get_vif_info(self._session, [ 586.805313] env[63372]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.805313] env[63372]: for vif in network_info: [ 586.805313] env[63372]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.805313] env[63372]: return self._sync_wrapper(fn, *args, **kwargs) [ 586.805313] env[63372]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.805313] env[63372]: self.wait() [ 586.805313] env[63372]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.805313] env[63372]: self[:] = self._gt.wait() [ 586.805313] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.805313] env[63372]: return self._exit_event.wait() [ 586.805313] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.805313] env[63372]: result = hub.switch() [ 586.805313] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.805313] env[63372]: return self.greenlet.switch() [ 586.805313] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.805313] env[63372]: result = function(*args, **kwargs) [ 586.806245] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.806245] env[63372]: return func(*args, **kwargs) [ 586.806245] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.806245] env[63372]: raise e [ 586.806245] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.806245] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 586.806245] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 586.806245] env[63372]: created_port_ids = self._update_ports_for_instance( [ 586.806245] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 586.806245] env[63372]: with excutils.save_and_reraise_exception(): [ 586.806245] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.806245] env[63372]: self.force_reraise() [ 586.806245] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.806245] env[63372]: raise self.value [ 586.806245] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 586.806245] env[63372]: updated_port = self._update_port( [ 586.806245] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.806245] env[63372]: _ensure_no_port_binding_failure(port) [ 586.806245] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.806245] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 586.806245] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 9eb8da80-70b2-45fe-92f5-022d9b9a217c, please check neutron logs for more information. [ 586.806245] env[63372]: During handling of the above exception, another exception occurred: [ 586.806245] env[63372]: Traceback (most recent call last): [ 586.806245] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2456, in _do_build_and_run_instance [ 586.806245] env[63372]: self._build_and_run_instance(context, instance, image, [ 586.806245] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2748, in _build_and_run_instance [ 586.806245] env[63372]: raise exception.RescheduledException( [ 586.806245] env[63372]: nova.exception.RescheduledException: Build of instance 7803b83f-2944-4e3b-87b1-0d3cea3455ad was re-scheduled: Binding failed for port 9eb8da80-70b2-45fe-92f5-022d9b9a217c, please check neutron logs for more information. [ 586.807371] env[63372]: During handling of the above exception, another exception occurred: [ 586.807371] env[63372]: Traceback (most recent call last): [ 586.807371] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 586.807371] env[63372]: func(*args, **kwargs) [ 586.807371] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.807371] env[63372]: return func(*args, **kwargs) [ 586.807371] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 586.807371] env[63372]: return f(*args, **kwargs) [ 586.807371] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2347, in _locked_do_build_and_run_instance [ 586.807371] env[63372]: result = self._do_build_and_run_instance(*args, **kwargs) [ 586.807371] env[63372]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 586.807371] env[63372]: with excutils.save_and_reraise_exception(): [ 586.807371] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.807371] env[63372]: self.force_reraise() [ 586.807371] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.807371] env[63372]: raise self.value [ 586.807371] env[63372]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 586.807371] env[63372]: return f(self, context, *args, **kw) [ 586.807371] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 586.807371] env[63372]: with excutils.save_and_reraise_exception(): [ 586.807371] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.807371] env[63372]: self.force_reraise() [ 586.807371] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.807371] env[63372]: raise self.value [ 586.807371] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 586.807371] env[63372]: return function(self, context, *args, **kwargs) [ 586.807371] env[63372]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 586.807371] env[63372]: return function(self, context, *args, **kwargs) [ 586.807371] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 586.807371] env[63372]: return function(self, context, *args, **kwargs) [ 586.807371] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2491, in _do_build_and_run_instance [ 586.807371] env[63372]: instance.save() [ 586.808437] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 586.808437] env[63372]: updates, result = self.indirection_api.object_action( [ 586.808437] env[63372]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 586.808437] env[63372]: return cctxt.call(context, 'object_action', objinst=objinst, [ 586.808437] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 586.808437] env[63372]: result = self.transport._send( [ 586.808437] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 586.808437] env[63372]: return self._driver.send(target, ctxt, message, [ 586.808437] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 586.808437] env[63372]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 586.808437] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 586.808437] env[63372]: raise result [ 586.808437] env[63372]: nova.exception_Remote.InstanceNotFound_Remote: Instance 7803b83f-2944-4e3b-87b1-0d3cea3455ad could not be found. [ 586.808437] env[63372]: Traceback (most recent call last): [ 586.808437] env[63372]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 586.808437] env[63372]: return getattr(target, method)(*args, **kwargs) [ 586.808437] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 586.808437] env[63372]: return fn(self, *args, **kwargs) [ 586.808437] env[63372]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 586.808437] env[63372]: old_ref, inst_ref = db.instance_update_and_get_original( [ 586.808437] env[63372]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 586.808437] env[63372]: return f(*args, **kwargs) [ 586.808437] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 586.808437] env[63372]: with excutils.save_and_reraise_exception() as ectxt: [ 586.808437] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.808437] env[63372]: self.force_reraise() [ 586.808437] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.808437] env[63372]: raise self.value [ 586.812126] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 586.812126] env[63372]: return f(*args, **kwargs) [ 586.812126] env[63372]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 586.812126] env[63372]: return f(context, *args, **kwargs) [ 586.812126] env[63372]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 586.812126] env[63372]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 586.812126] env[63372]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 586.812126] env[63372]: raise exception.InstanceNotFound(instance_id=uuid) [ 586.812126] env[63372]: nova.exception.InstanceNotFound: Instance 7803b83f-2944-4e3b-87b1-0d3cea3455ad could not be found. [ 587.080967] env[63372]: DEBUG nova.compute.manager [req-32c07d4d-1b8c-4589-a960-95ac19cb82dd req-fbcb08c9-cf92-47c1-81d2-f0349b39e675 service nova] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Received event network-vif-deleted-7c7784d9-31f5-4c0a-b5d8-93d147d846c0 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 587.168928] env[63372]: DEBUG nova.network.neutron [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.308040] env[63372]: DEBUG nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 587.343916] env[63372]: DEBUG nova.network.neutron [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.468320] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf31da21-0a91-4525-9520-3d951ffe3f2c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.479442] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19dbbeda-0f8b-46e9-8bac-5264a7e9241b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.517690] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a7b2df-30c5-4e5c-bb9a-520f51b3a390 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.526109] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a531ec68-1cdb-4961-989e-27b5e468271b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.542146] env[63372]: DEBUG nova.compute.provider_tree [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.624024] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Acquiring lock "276c3490-240a-499b-ba64-4b02510b169c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.624274] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Lock "276c3490-240a-499b-ba64-4b02510b169c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.832846] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.843312] env[63372]: INFO nova.scheduler.client.report [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Deleted allocations for instance 9ec742c1-894f-4baa-a451-01ebb4d2816a [ 587.851145] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Releasing lock "refresh_cache-dc9c4a56-e0af-473b-b47b-2061fe89c6d6" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.851379] env[63372]: DEBUG nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 587.851597] env[63372]: DEBUG nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.855039] env[63372]: DEBUG nova.network.neutron [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 587.885339] env[63372]: DEBUG nova.network.neutron [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.045445] env[63372]: DEBUG nova.scheduler.client.report [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 588.358936] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f0073011-9afe-44ae-8930-1a4e25536362 tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Lock "9ec742c1-894f-4baa-a451-01ebb4d2816a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.606s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.388750] env[63372]: DEBUG nova.network.neutron [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.553039] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.018s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.553706] env[63372]: ERROR nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 92a840c0-9f37-4af2-9b74-f20b406f353d, please check neutron logs for more information. [ 588.553706] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Traceback (most recent call last): [ 588.553706] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.553706] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] self.driver.spawn(context, instance, image_meta, [ 588.553706] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 588.553706] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.553706] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.553706] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] vm_ref = self.build_virtual_machine(instance, [ 588.553706] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.553706] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.553706] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.554150] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] for vif in network_info: [ 588.554150] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.554150] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] return self._sync_wrapper(fn, *args, **kwargs) [ 588.554150] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.554150] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] self.wait() [ 588.554150] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.554150] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] self[:] = self._gt.wait() [ 588.554150] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.554150] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] return self._exit_event.wait() [ 588.554150] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.554150] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] result = hub.switch() [ 588.554150] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.554150] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] return self.greenlet.switch() [ 588.554663] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.554663] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] result = function(*args, **kwargs) [ 588.554663] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.554663] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] return func(*args, **kwargs) [ 588.554663] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.554663] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] raise e [ 588.554663] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.554663] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] nwinfo = self.network_api.allocate_for_instance( [ 588.554663] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 588.554663] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] created_port_ids = self._update_ports_for_instance( [ 588.554663] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 588.554663] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] with excutils.save_and_reraise_exception(): [ 588.554663] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.555145] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] self.force_reraise() [ 588.555145] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.555145] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] raise self.value [ 588.555145] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 588.555145] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] updated_port = self._update_port( [ 588.555145] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.555145] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] _ensure_no_port_binding_failure(port) [ 588.555145] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.555145] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] raise exception.PortBindingFailed(port_id=port['id']) [ 588.555145] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] nova.exception.PortBindingFailed: Binding failed for port 92a840c0-9f37-4af2-9b74-f20b406f353d, please check neutron logs for more information. [ 588.555145] env[63372]: ERROR nova.compute.manager [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] [ 588.555471] env[63372]: DEBUG nova.compute.utils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Binding failed for port 92a840c0-9f37-4af2-9b74-f20b406f353d, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 588.561270] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.095s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.563908] env[63372]: DEBUG nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Build of instance 392f418f-c8c7-41e3-9bb6-f68e1cab75f3 was re-scheduled: Binding failed for port 92a840c0-9f37-4af2-9b74-f20b406f353d, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 588.564687] env[63372]: DEBUG nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 588.564687] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Acquiring lock "refresh_cache-392f418f-c8c7-41e3-9bb6-f68e1cab75f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.564853] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Acquired lock "refresh_cache-392f418f-c8c7-41e3-9bb6-f68e1cab75f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.564887] env[63372]: DEBUG nova.network.neutron [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 588.862286] env[63372]: DEBUG nova.compute.manager [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 588.891615] env[63372]: INFO nova.compute.manager [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] [instance: dc9c4a56-e0af-473b-b47b-2061fe89c6d6] Took 1.04 seconds to deallocate network for instance. [ 589.104263] env[63372]: DEBUG nova.network.neutron [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.312104] env[63372]: DEBUG nova.network.neutron [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.387531] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.518226] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b362be4-6ce7-4af4-abb6-1ad1f6d5376d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.527082] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e47eb9f7-11d8-40b6-9244-ff4cb053deff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.570414] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e604cbf7-98c7-42c4-a197-9beee10e9d8b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.580487] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d63e6e-ac76-40b8-b87f-5ec1aa9e40c1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.595405] env[63372]: DEBUG nova.compute.provider_tree [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.814952] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Releasing lock "refresh_cache-392f418f-c8c7-41e3-9bb6-f68e1cab75f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.814952] env[63372]: DEBUG nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 589.815303] env[63372]: DEBUG nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.815395] env[63372]: DEBUG nova.network.neutron [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 589.849462] env[63372]: DEBUG nova.network.neutron [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.958304] env[63372]: INFO nova.scheduler.client.report [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Deleted allocations for instance dc9c4a56-e0af-473b-b47b-2061fe89c6d6 [ 590.098866] env[63372]: DEBUG nova.scheduler.client.report [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.354466] env[63372]: DEBUG nova.network.neutron [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.476416] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b244845d-75db-4dc1-b770-5eddc8b9badd tempest-ServersAdminTestJSON-1559128535 tempest-ServersAdminTestJSON-1559128535-project-member] Lock "dc9c4a56-e0af-473b-b47b-2061fe89c6d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.594s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.605967] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.045s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.606937] env[63372]: ERROR nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 64490f1e-3e00-4fc9-92fb-a2e454998270, please check neutron logs for more information. [ 590.606937] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Traceback (most recent call last): [ 590.606937] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 590.606937] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] self.driver.spawn(context, instance, image_meta, [ 590.606937] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 590.606937] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] self._vmops.spawn(context, instance, image_meta, injected_files, [ 590.606937] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 590.606937] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] vm_ref = self.build_virtual_machine(instance, [ 590.606937] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 590.606937] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] vif_infos = vmwarevif.get_vif_info(self._session, [ 590.606937] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 590.607351] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] for vif in network_info: [ 590.607351] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 590.607351] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] return self._sync_wrapper(fn, *args, **kwargs) [ 590.607351] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 590.607351] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] self.wait() [ 590.607351] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 590.607351] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] self[:] = self._gt.wait() [ 590.607351] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 590.607351] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] return self._exit_event.wait() [ 590.607351] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 590.607351] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] result = hub.switch() [ 590.607351] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 590.607351] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] return self.greenlet.switch() [ 590.607706] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.607706] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] result = function(*args, **kwargs) [ 590.607706] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 590.607706] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] return func(*args, **kwargs) [ 590.607706] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.607706] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] raise e [ 590.607706] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.607706] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] nwinfo = self.network_api.allocate_for_instance( [ 590.607706] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 590.607706] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] created_port_ids = self._update_ports_for_instance( [ 590.607706] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 590.607706] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] with excutils.save_and_reraise_exception(): [ 590.607706] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.608069] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] self.force_reraise() [ 590.608069] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.608069] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] raise self.value [ 590.608069] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 590.608069] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] updated_port = self._update_port( [ 590.608069] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.608069] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] _ensure_no_port_binding_failure(port) [ 590.608069] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.608069] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] raise exception.PortBindingFailed(port_id=port['id']) [ 590.608069] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] nova.exception.PortBindingFailed: Binding failed for port 64490f1e-3e00-4fc9-92fb-a2e454998270, please check neutron logs for more information. [ 590.608069] env[63372]: ERROR nova.compute.manager [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] [ 590.608378] env[63372]: DEBUG nova.compute.utils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Binding failed for port 64490f1e-3e00-4fc9-92fb-a2e454998270, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 590.612628] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.256s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.614171] env[63372]: INFO nova.compute.claims [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 590.619417] env[63372]: DEBUG nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Build of instance 9d52ebd0-f38e-4825-a974-f9a0c0643116 was re-scheduled: Binding failed for port 64490f1e-3e00-4fc9-92fb-a2e454998270, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 590.620379] env[63372]: DEBUG nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 590.620715] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "refresh_cache-9d52ebd0-f38e-4825-a974-f9a0c0643116" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.620768] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquired lock "refresh_cache-9d52ebd0-f38e-4825-a974-f9a0c0643116" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.620925] env[63372]: DEBUG nova.network.neutron [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 590.861898] env[63372]: INFO nova.compute.manager [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] [instance: 392f418f-c8c7-41e3-9bb6-f68e1cab75f3] Took 1.04 seconds to deallocate network for instance. [ 590.981413] env[63372]: DEBUG nova.compute.manager [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.151360] env[63372]: DEBUG nova.network.neutron [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.286406] env[63372]: DEBUG nova.network.neutron [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.517674] env[63372]: DEBUG oslo_concurrency.lockutils [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.789989] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Releasing lock "refresh_cache-9d52ebd0-f38e-4825-a974-f9a0c0643116" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.790381] env[63372]: DEBUG nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 591.790595] env[63372]: DEBUG nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 591.790786] env[63372]: DEBUG nova.network.neutron [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 591.811327] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquiring lock "d50919ac-3a0b-46ac-a837-ca3e6ad05173" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.811563] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Lock "d50919ac-3a0b-46ac-a837-ca3e6ad05173" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.819406] env[63372]: DEBUG nova.network.neutron [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.901232] env[63372]: INFO nova.scheduler.client.report [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Deleted allocations for instance 392f418f-c8c7-41e3-9bb6-f68e1cab75f3 [ 592.038215] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a3e320-14e1-4720-9df1-d8240cbc5407 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.045916] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be632a18-4e42-476f-ac29-72af03b1bed1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.079555] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-297951d6-1db5-4327-8bcd-70f9358f63be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.087196] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1ec5c7-acd2-4f8c-80ef-b852cadd183a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.102974] env[63372]: DEBUG nova.compute.provider_tree [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.322670] env[63372]: DEBUG nova.network.neutron [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.410978] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1800bd2-2ab4-4e14-a7ee-655fdabe43ba tempest-VolumesAssistedSnapshotsTest-668722185 tempest-VolumesAssistedSnapshotsTest-668722185-project-member] Lock "392f418f-c8c7-41e3-9bb6-f68e1cab75f3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.918s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.606841] env[63372]: DEBUG nova.scheduler.client.report [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.826117] env[63372]: INFO nova.compute.manager [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: 9d52ebd0-f38e-4825-a974-f9a0c0643116] Took 1.04 seconds to deallocate network for instance. [ 592.914622] env[63372]: DEBUG nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 593.119150] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.119150] env[63372]: DEBUG nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 593.122195] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.758s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.124666] env[63372]: INFO nova.compute.claims [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.397302] env[63372]: DEBUG oslo_concurrency.lockutils [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Acquiring lock "44fba254-7dec-4458-9ae3-fdbbe4895de5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.397524] env[63372]: DEBUG oslo_concurrency.lockutils [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Lock "44fba254-7dec-4458-9ae3-fdbbe4895de5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.443095] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.631500] env[63372]: DEBUG nova.compute.utils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 593.640922] env[63372]: DEBUG nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 593.641124] env[63372]: DEBUG nova.network.neutron [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 593.734386] env[63372]: DEBUG nova.policy [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4bb27c3911354fec9de21a10066db1fc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c5c6449b70e94bdc861855bd92d38c28', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 593.866194] env[63372]: INFO nova.scheduler.client.report [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Deleted allocations for instance 9d52ebd0-f38e-4825-a974-f9a0c0643116 [ 594.142548] env[63372]: DEBUG nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 594.380194] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0cf1657e-bd22-4709-a1b8-1978b60d4e63 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "9d52ebd0-f38e-4825-a974-f9a0c0643116" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.958s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.487072] env[63372]: DEBUG nova.network.neutron [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Successfully created port: 31028e9f-4749-4a6a-815e-fb2daf8f6c84 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 594.556339] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b535d0b-a1e3-4d8d-bd8d-492dd88d86b8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.567426] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a7ace2-8c37-48d8-b915-fed729e123cf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.602266] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1715e6f-be28-43e3-88ff-78f3d75918be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.610536] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96243aca-b60c-4c56-ac4e-1fa3e521ff75 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.628038] env[63372]: DEBUG nova.compute.provider_tree [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.881171] env[63372]: DEBUG nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 595.133021] env[63372]: DEBUG nova.scheduler.client.report [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.156579] env[63372]: DEBUG nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 595.182321] env[63372]: DEBUG nova.virt.hardware [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 595.182321] env[63372]: DEBUG nova.virt.hardware [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 595.182321] env[63372]: DEBUG nova.virt.hardware [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 595.182321] env[63372]: DEBUG nova.virt.hardware [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 595.182804] env[63372]: DEBUG nova.virt.hardware [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 595.182804] env[63372]: DEBUG nova.virt.hardware [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 595.182804] env[63372]: DEBUG nova.virt.hardware [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 595.182804] env[63372]: DEBUG nova.virt.hardware [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 595.182945] env[63372]: DEBUG nova.virt.hardware [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 595.183369] env[63372]: DEBUG nova.virt.hardware [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 595.183556] env[63372]: DEBUG nova.virt.hardware [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 595.184691] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89c111e-9d4a-4ab8-9445-07ca04295f8b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.197789] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9558eabb-b7b7-4c47-8a5c-e62836095799 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.302797] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Acquiring lock "85883c62-6d5e-4fa6-b4f0-afac984dc6bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.303252] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Lock "85883c62-6d5e-4fa6-b4f0-afac984dc6bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.408955] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.636700] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.637236] env[63372]: DEBUG nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.639788] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.169s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.641369] env[63372]: INFO nova.compute.claims [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.132105] env[63372]: DEBUG nova.compute.manager [req-4e9b98e4-c31c-4942-90fe-2a4696b5062d req-1835410c-d660-482c-9e16-0fcaa4788d82 service nova] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Received event network-changed-31028e9f-4749-4a6a-815e-fb2daf8f6c84 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.132490] env[63372]: DEBUG nova.compute.manager [req-4e9b98e4-c31c-4942-90fe-2a4696b5062d req-1835410c-d660-482c-9e16-0fcaa4788d82 service nova] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Refreshing instance network info cache due to event network-changed-31028e9f-4749-4a6a-815e-fb2daf8f6c84. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 596.132826] env[63372]: DEBUG oslo_concurrency.lockutils [req-4e9b98e4-c31c-4942-90fe-2a4696b5062d req-1835410c-d660-482c-9e16-0fcaa4788d82 service nova] Acquiring lock "refresh_cache-fe96690d-4ca0-413b-9e04-09e7f73a0bb7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.132826] env[63372]: DEBUG oslo_concurrency.lockutils [req-4e9b98e4-c31c-4942-90fe-2a4696b5062d req-1835410c-d660-482c-9e16-0fcaa4788d82 service nova] Acquired lock "refresh_cache-fe96690d-4ca0-413b-9e04-09e7f73a0bb7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.133014] env[63372]: DEBUG nova.network.neutron [req-4e9b98e4-c31c-4942-90fe-2a4696b5062d req-1835410c-d660-482c-9e16-0fcaa4788d82 service nova] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Refreshing network info cache for port 31028e9f-4749-4a6a-815e-fb2daf8f6c84 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 596.146121] env[63372]: DEBUG nova.compute.utils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.152067] env[63372]: DEBUG nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 596.152067] env[63372]: DEBUG nova.network.neutron [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 596.358111] env[63372]: DEBUG nova.policy [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1048d032dc00441f9c2f99b60f357dfb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5cf13f5cc32c4b3ebaee47f03bb64ba7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 596.497897] env[63372]: ERROR nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 31028e9f-4749-4a6a-815e-fb2daf8f6c84, please check neutron logs for more information. [ 596.497897] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 596.497897] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.497897] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 596.497897] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.497897] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 596.497897] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.497897] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 596.497897] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.497897] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 596.497897] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.497897] env[63372]: ERROR nova.compute.manager raise self.value [ 596.497897] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.497897] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 596.497897] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.497897] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 596.498491] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.498491] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 596.498491] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 31028e9f-4749-4a6a-815e-fb2daf8f6c84, please check neutron logs for more information. [ 596.498491] env[63372]: ERROR nova.compute.manager [ 596.498491] env[63372]: Traceback (most recent call last): [ 596.498491] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 596.498491] env[63372]: listener.cb(fileno) [ 596.498491] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.498491] env[63372]: result = function(*args, **kwargs) [ 596.498491] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.498491] env[63372]: return func(*args, **kwargs) [ 596.498491] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.498491] env[63372]: raise e [ 596.498491] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.498491] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 596.498491] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.498491] env[63372]: created_port_ids = self._update_ports_for_instance( [ 596.498491] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.498491] env[63372]: with excutils.save_and_reraise_exception(): [ 596.498491] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.498491] env[63372]: self.force_reraise() [ 596.498491] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.498491] env[63372]: raise self.value [ 596.498491] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.498491] env[63372]: updated_port = self._update_port( [ 596.498491] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.498491] env[63372]: _ensure_no_port_binding_failure(port) [ 596.498491] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.498491] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 596.499339] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 31028e9f-4749-4a6a-815e-fb2daf8f6c84, please check neutron logs for more information. [ 596.499339] env[63372]: Removing descriptor: 19 [ 596.499339] env[63372]: ERROR nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 31028e9f-4749-4a6a-815e-fb2daf8f6c84, please check neutron logs for more information. [ 596.499339] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Traceback (most recent call last): [ 596.499339] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 596.499339] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] yield resources [ 596.499339] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.499339] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] self.driver.spawn(context, instance, image_meta, [ 596.499339] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 596.499339] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.499339] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.499339] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] vm_ref = self.build_virtual_machine(instance, [ 596.499716] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.499716] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.499716] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.499716] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] for vif in network_info: [ 596.499716] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.499716] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] return self._sync_wrapper(fn, *args, **kwargs) [ 596.499716] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.499716] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] self.wait() [ 596.499716] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.499716] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] self[:] = self._gt.wait() [ 596.499716] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.499716] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] return self._exit_event.wait() [ 596.499716] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.500178] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] result = hub.switch() [ 596.500178] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.500178] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] return self.greenlet.switch() [ 596.500178] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.500178] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] result = function(*args, **kwargs) [ 596.500178] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.500178] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] return func(*args, **kwargs) [ 596.500178] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.500178] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] raise e [ 596.500178] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.500178] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] nwinfo = self.network_api.allocate_for_instance( [ 596.500178] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.500178] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] created_port_ids = self._update_ports_for_instance( [ 596.500578] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.500578] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] with excutils.save_and_reraise_exception(): [ 596.500578] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.500578] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] self.force_reraise() [ 596.500578] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.500578] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] raise self.value [ 596.500578] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.500578] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] updated_port = self._update_port( [ 596.500578] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.500578] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] _ensure_no_port_binding_failure(port) [ 596.500578] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.500578] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] raise exception.PortBindingFailed(port_id=port['id']) [ 596.501158] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] nova.exception.PortBindingFailed: Binding failed for port 31028e9f-4749-4a6a-815e-fb2daf8f6c84, please check neutron logs for more information. [ 596.501158] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] [ 596.501158] env[63372]: INFO nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Terminating instance [ 596.501975] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Acquiring lock "refresh_cache-fe96690d-4ca0-413b-9e04-09e7f73a0bb7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.654791] env[63372]: DEBUG nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.662548] env[63372]: DEBUG nova.network.neutron [req-4e9b98e4-c31c-4942-90fe-2a4696b5062d req-1835410c-d660-482c-9e16-0fcaa4788d82 service nova] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.775771] env[63372]: DEBUG nova.network.neutron [req-4e9b98e4-c31c-4942-90fe-2a4696b5062d req-1835410c-d660-482c-9e16-0fcaa4788d82 service nova] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.045231] env[63372]: DEBUG nova.network.neutron [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Successfully created port: 5f1c17ad-7301-48b8-8a7d-aacd4717dc74 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.056008] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13c81f6a-8d73-4cd8-82e8-93cc178cc7e0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.064795] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66de2579-c509-4373-9130-f52f0c258a3a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.107500] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf93e5d6-7c58-4994-a349-87e9aaf92d3d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.118178] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b28cdf33-45f7-4fdb-aad5-b3bbdf986d6b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.133594] env[63372]: DEBUG nova.compute.provider_tree [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.280764] env[63372]: DEBUG oslo_concurrency.lockutils [req-4e9b98e4-c31c-4942-90fe-2a4696b5062d req-1835410c-d660-482c-9e16-0fcaa4788d82 service nova] Releasing lock "refresh_cache-fe96690d-4ca0-413b-9e04-09e7f73a0bb7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.281240] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Acquired lock "refresh_cache-fe96690d-4ca0-413b-9e04-09e7f73a0bb7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.281475] env[63372]: DEBUG nova.network.neutron [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 597.636720] env[63372]: DEBUG nova.scheduler.client.report [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.664922] env[63372]: DEBUG nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.696425] env[63372]: DEBUG nova.virt.hardware [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.696757] env[63372]: DEBUG nova.virt.hardware [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.696942] env[63372]: DEBUG nova.virt.hardware [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.697219] env[63372]: DEBUG nova.virt.hardware [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.697374] env[63372]: DEBUG nova.virt.hardware [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.697951] env[63372]: DEBUG nova.virt.hardware [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.698224] env[63372]: DEBUG nova.virt.hardware [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.698388] env[63372]: DEBUG nova.virt.hardware [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.698552] env[63372]: DEBUG nova.virt.hardware [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.698707] env[63372]: DEBUG nova.virt.hardware [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.698900] env[63372]: DEBUG nova.virt.hardware [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.700841] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4263cd7e-b114-4abc-b27d-36e4aac39be2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.709876] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9edc833-ff4e-4cd0-8e62-7c5c0ead3022 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.815113] env[63372]: DEBUG nova.network.neutron [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.078770] env[63372]: DEBUG nova.network.neutron [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.144684] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.505s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.145303] env[63372]: DEBUG nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 598.149211] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.605s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.150745] env[63372]: INFO nova.compute.claims [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 598.212991] env[63372]: DEBUG nova.compute.manager [req-14d8cc85-ebe5-4db4-abcc-8c183e9ed4f2 req-4ff781e8-0826-4a1d-b10f-2fd3c158864a service nova] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Received event network-vif-deleted-31028e9f-4749-4a6a-815e-fb2daf8f6c84 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 598.407234] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 598.408047] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 598.569385] env[63372]: ERROR nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5f1c17ad-7301-48b8-8a7d-aacd4717dc74, please check neutron logs for more information. [ 598.569385] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 598.569385] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.569385] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 598.569385] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.569385] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 598.569385] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.569385] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 598.569385] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.569385] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 598.569385] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.569385] env[63372]: ERROR nova.compute.manager raise self.value [ 598.569385] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.569385] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 598.569385] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.569385] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 598.569874] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.569874] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 598.569874] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5f1c17ad-7301-48b8-8a7d-aacd4717dc74, please check neutron logs for more information. [ 598.569874] env[63372]: ERROR nova.compute.manager [ 598.569874] env[63372]: Traceback (most recent call last): [ 598.569874] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 598.569874] env[63372]: listener.cb(fileno) [ 598.569874] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.569874] env[63372]: result = function(*args, **kwargs) [ 598.569874] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.569874] env[63372]: return func(*args, **kwargs) [ 598.569874] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.569874] env[63372]: raise e [ 598.569874] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.569874] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 598.569874] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.569874] env[63372]: created_port_ids = self._update_ports_for_instance( [ 598.569874] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.569874] env[63372]: with excutils.save_and_reraise_exception(): [ 598.569874] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.569874] env[63372]: self.force_reraise() [ 598.569874] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.569874] env[63372]: raise self.value [ 598.569874] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.569874] env[63372]: updated_port = self._update_port( [ 598.569874] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.569874] env[63372]: _ensure_no_port_binding_failure(port) [ 598.569874] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.569874] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 598.571295] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 5f1c17ad-7301-48b8-8a7d-aacd4717dc74, please check neutron logs for more information. [ 598.571295] env[63372]: Removing descriptor: 14 [ 598.571295] env[63372]: ERROR nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5f1c17ad-7301-48b8-8a7d-aacd4717dc74, please check neutron logs for more information. [ 598.571295] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Traceback (most recent call last): [ 598.571295] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 598.571295] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] yield resources [ 598.571295] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 598.571295] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] self.driver.spawn(context, instance, image_meta, [ 598.571295] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 598.571295] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.571295] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.571295] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] vm_ref = self.build_virtual_machine(instance, [ 598.571853] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.571853] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.571853] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.571853] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] for vif in network_info: [ 598.571853] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.571853] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] return self._sync_wrapper(fn, *args, **kwargs) [ 598.571853] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.571853] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] self.wait() [ 598.571853] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.571853] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] self[:] = self._gt.wait() [ 598.571853] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.571853] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] return self._exit_event.wait() [ 598.571853] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.572430] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] result = hub.switch() [ 598.572430] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.572430] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] return self.greenlet.switch() [ 598.572430] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.572430] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] result = function(*args, **kwargs) [ 598.572430] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.572430] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] return func(*args, **kwargs) [ 598.572430] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.572430] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] raise e [ 598.572430] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.572430] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] nwinfo = self.network_api.allocate_for_instance( [ 598.572430] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.572430] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] created_port_ids = self._update_ports_for_instance( [ 598.573076] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.573076] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] with excutils.save_and_reraise_exception(): [ 598.573076] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.573076] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] self.force_reraise() [ 598.573076] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.573076] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] raise self.value [ 598.573076] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.573076] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] updated_port = self._update_port( [ 598.573076] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.573076] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] _ensure_no_port_binding_failure(port) [ 598.573076] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.573076] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] raise exception.PortBindingFailed(port_id=port['id']) [ 598.573630] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] nova.exception.PortBindingFailed: Binding failed for port 5f1c17ad-7301-48b8-8a7d-aacd4717dc74, please check neutron logs for more information. [ 598.573630] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] [ 598.573630] env[63372]: INFO nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Terminating instance [ 598.573630] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "refresh_cache-d0310e79-e07a-4581-8cd8-34437cd9ea7d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.573630] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquired lock "refresh_cache-d0310e79-e07a-4581-8cd8-34437cd9ea7d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.573630] env[63372]: DEBUG nova.network.neutron [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 598.582119] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Releasing lock "refresh_cache-fe96690d-4ca0-413b-9e04-09e7f73a0bb7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.582656] env[63372]: DEBUG nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 598.582854] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 598.583173] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f97495b-122d-4962-92c1-e0914410311b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.598900] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c76da60-39bf-4f23-b098-9b027c633311 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.635357] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fe96690d-4ca0-413b-9e04-09e7f73a0bb7 could not be found. [ 598.635936] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 598.635936] env[63372]: INFO nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Took 0.05 seconds to destroy the instance on the hypervisor. [ 598.636163] env[63372]: DEBUG oslo.service.loopingcall [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 598.636793] env[63372]: DEBUG nova.compute.manager [-] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.637105] env[63372]: DEBUG nova.network.neutron [-] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 598.655585] env[63372]: DEBUG nova.compute.utils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 598.671230] env[63372]: DEBUG nova.network.neutron [-] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.672373] env[63372]: DEBUG nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 598.672558] env[63372]: DEBUG nova.network.neutron [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 598.759388] env[63372]: DEBUG nova.policy [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2b1aa4bdcd84be08082e5c7417eec43', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8a6401c846184105864d27263d561660', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 598.917617] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 598.917906] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Starting heal instance info cache {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 598.918445] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Rebuilding the list of instances to heal {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 599.107836] env[63372]: DEBUG nova.network.neutron [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.170887] env[63372]: DEBUG nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 599.181397] env[63372]: DEBUG nova.network.neutron [-] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.224230] env[63372]: DEBUG nova.network.neutron [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.368366] env[63372]: DEBUG nova.network.neutron [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Successfully created port: 852395e7-45d6-4d16-8b05-2144329e6dc4 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 599.420979] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 599.421165] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 599.421296] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 599.421419] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 599.421541] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 599.421696] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Didn't find any instances for network info cache update. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 599.421910] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.422048] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.422196] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.422338] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.422474] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.422624] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.422749] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63372) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 599.422955] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager.update_available_resource {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.558948] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4993134b-2b30-4962-b94d-4630ffb2de9b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.572667] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e45a77-b9c3-4f64-8550-a4116a4c36d6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.605745] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7760e68a-0d64-4fd9-8ef7-82314292ea30 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.614446] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243d0e27-ee88-421b-8ef3-e9ddc628640d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.632055] env[63372]: DEBUG nova.compute.provider_tree [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.689273] env[63372]: INFO nova.compute.manager [-] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Took 1.05 seconds to deallocate network for instance. [ 599.693814] env[63372]: DEBUG nova.compute.claims [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 599.693814] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.732661] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Releasing lock "refresh_cache-d0310e79-e07a-4581-8cd8-34437cd9ea7d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.736205] env[63372]: DEBUG nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.736205] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 599.736205] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc167557-72e9-4aea-beba-daa0e695502b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.745543] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d740d7-b770-463d-b2ba-b5d151bd55ec {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.776371] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d0310e79-e07a-4581-8cd8-34437cd9ea7d could not be found. [ 599.776371] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 599.776641] env[63372]: INFO nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 599.777192] env[63372]: DEBUG oslo.service.loopingcall [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.777192] env[63372]: DEBUG nova.compute.manager [-] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.777254] env[63372]: DEBUG nova.network.neutron [-] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.794790] env[63372]: DEBUG nova.network.neutron [-] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.925989] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.135056] env[63372]: DEBUG nova.scheduler.client.report [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.182720] env[63372]: DEBUG nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 600.216139] env[63372]: DEBUG nova.virt.hardware [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:34:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1449644025',id=23,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-941503515',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 600.216463] env[63372]: DEBUG nova.virt.hardware [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 600.216623] env[63372]: DEBUG nova.virt.hardware [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 600.217041] env[63372]: DEBUG nova.virt.hardware [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 600.217218] env[63372]: DEBUG nova.virt.hardware [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 600.217367] env[63372]: DEBUG nova.virt.hardware [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 600.217575] env[63372]: DEBUG nova.virt.hardware [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 600.217729] env[63372]: DEBUG nova.virt.hardware [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 600.217891] env[63372]: DEBUG nova.virt.hardware [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 600.218061] env[63372]: DEBUG nova.virt.hardware [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 600.218272] env[63372]: DEBUG nova.virt.hardware [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 600.219466] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b68ce7be-b1ff-4463-ad30-258bc96483df {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.228977] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472a63ee-a039-4c11-b096-d0fd867c8a27 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.297064] env[63372]: DEBUG nova.network.neutron [-] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.402485] env[63372]: ERROR nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 852395e7-45d6-4d16-8b05-2144329e6dc4, please check neutron logs for more information. [ 600.402485] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 600.402485] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.402485] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 600.402485] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.402485] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 600.402485] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.402485] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 600.402485] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.402485] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 600.402485] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.402485] env[63372]: ERROR nova.compute.manager raise self.value [ 600.402485] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.402485] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 600.402485] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.402485] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 600.402860] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.402860] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 600.402860] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 852395e7-45d6-4d16-8b05-2144329e6dc4, please check neutron logs for more information. [ 600.402860] env[63372]: ERROR nova.compute.manager [ 600.402860] env[63372]: Traceback (most recent call last): [ 600.402860] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 600.402860] env[63372]: listener.cb(fileno) [ 600.402860] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.402860] env[63372]: result = function(*args, **kwargs) [ 600.402860] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.402860] env[63372]: return func(*args, **kwargs) [ 600.402860] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.402860] env[63372]: raise e [ 600.402860] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.402860] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 600.402860] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.402860] env[63372]: created_port_ids = self._update_ports_for_instance( [ 600.402860] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.402860] env[63372]: with excutils.save_and_reraise_exception(): [ 600.402860] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.402860] env[63372]: self.force_reraise() [ 600.402860] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.402860] env[63372]: raise self.value [ 600.402860] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.402860] env[63372]: updated_port = self._update_port( [ 600.402860] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.402860] env[63372]: _ensure_no_port_binding_failure(port) [ 600.402860] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.402860] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 600.403513] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 852395e7-45d6-4d16-8b05-2144329e6dc4, please check neutron logs for more information. [ 600.403513] env[63372]: Removing descriptor: 14 [ 600.403513] env[63372]: ERROR nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 852395e7-45d6-4d16-8b05-2144329e6dc4, please check neutron logs for more information. [ 600.403513] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Traceback (most recent call last): [ 600.403513] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 600.403513] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] yield resources [ 600.403513] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.403513] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] self.driver.spawn(context, instance, image_meta, [ 600.403513] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 600.403513] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.403513] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.403513] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] vm_ref = self.build_virtual_machine(instance, [ 600.403768] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.403768] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.403768] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.403768] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] for vif in network_info: [ 600.403768] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.403768] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] return self._sync_wrapper(fn, *args, **kwargs) [ 600.403768] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.403768] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] self.wait() [ 600.403768] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.403768] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] self[:] = self._gt.wait() [ 600.403768] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.403768] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] return self._exit_event.wait() [ 600.403768] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.404028] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] result = hub.switch() [ 600.404028] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.404028] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] return self.greenlet.switch() [ 600.404028] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.404028] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] result = function(*args, **kwargs) [ 600.404028] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.404028] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] return func(*args, **kwargs) [ 600.404028] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.404028] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] raise e [ 600.404028] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.404028] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] nwinfo = self.network_api.allocate_for_instance( [ 600.404028] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.404028] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] created_port_ids = self._update_ports_for_instance( [ 600.404339] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.404339] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] with excutils.save_and_reraise_exception(): [ 600.404339] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.404339] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] self.force_reraise() [ 600.404339] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.404339] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] raise self.value [ 600.404339] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.404339] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] updated_port = self._update_port( [ 600.404339] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.404339] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] _ensure_no_port_binding_failure(port) [ 600.404339] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.404339] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] raise exception.PortBindingFailed(port_id=port['id']) [ 600.404629] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] nova.exception.PortBindingFailed: Binding failed for port 852395e7-45d6-4d16-8b05-2144329e6dc4, please check neutron logs for more information. [ 600.404629] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] [ 600.404629] env[63372]: INFO nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Terminating instance [ 600.406120] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Acquiring lock "refresh_cache-6a6268e4-0539-45cf-9335-cde31473d1f9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.406606] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Acquired lock "refresh_cache-6a6268e4-0539-45cf-9335-cde31473d1f9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.406797] env[63372]: DEBUG nova.network.neutron [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 600.629300] env[63372]: DEBUG nova.compute.manager [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Received event network-changed-5f1c17ad-7301-48b8-8a7d-aacd4717dc74 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.629513] env[63372]: DEBUG nova.compute.manager [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Refreshing instance network info cache due to event network-changed-5f1c17ad-7301-48b8-8a7d-aacd4717dc74. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 600.629718] env[63372]: DEBUG oslo_concurrency.lockutils [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] Acquiring lock "refresh_cache-d0310e79-e07a-4581-8cd8-34437cd9ea7d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.629855] env[63372]: DEBUG oslo_concurrency.lockutils [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] Acquired lock "refresh_cache-d0310e79-e07a-4581-8cd8-34437cd9ea7d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.630017] env[63372]: DEBUG nova.network.neutron [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Refreshing network info cache for port 5f1c17ad-7301-48b8-8a7d-aacd4717dc74 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 600.642632] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.643068] env[63372]: DEBUG nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.647793] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.221s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.799784] env[63372]: INFO nova.compute.manager [-] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Took 1.02 seconds to deallocate network for instance. [ 600.802800] env[63372]: DEBUG nova.compute.claims [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.803148] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.935906] env[63372]: DEBUG nova.network.neutron [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.101026] env[63372]: DEBUG nova.network.neutron [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.158195] env[63372]: DEBUG nova.compute.utils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 601.166559] env[63372]: DEBUG nova.network.neutron [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.167354] env[63372]: DEBUG nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 601.167631] env[63372]: DEBUG nova.network.neutron [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 601.261747] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.262348] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.464785] env[63372]: DEBUG nova.policy [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '988e5c6fbca24550ac11c641a9e8a9e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9f2294fd9854bb0ba7074a85894b0b8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 601.472690] env[63372]: DEBUG nova.network.neutron [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.581663] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cac0cec-9469-4e73-a8c3-688720862749 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.590976] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39526e7c-a643-42b4-9f06-af0cfade108f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.631162] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Releasing lock "refresh_cache-6a6268e4-0539-45cf-9335-cde31473d1f9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.631568] env[63372]: DEBUG nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 601.631754] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 601.632660] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-65f54e1b-ae5f-43a4-8b2d-aa8ba570136a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.635064] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0294175-e975-47e9-831e-62dfacdaab6f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.644832] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70e0ace-14b8-4f2e-becc-d8492a4a8786 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.650986] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b819603b-baa0-49d4-83ff-885dc77fbd27 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.667097] env[63372]: DEBUG nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.677135] env[63372]: DEBUG nova.compute.provider_tree [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.683908] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6a6268e4-0539-45cf-9335-cde31473d1f9 could not be found. [ 601.683908] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 601.683908] env[63372]: INFO nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 601.683908] env[63372]: DEBUG oslo.service.loopingcall [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 601.683908] env[63372]: DEBUG nova.compute.manager [-] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.684095] env[63372]: DEBUG nova.network.neutron [-] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 601.724908] env[63372]: DEBUG nova.network.neutron [-] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.976908] env[63372]: DEBUG oslo_concurrency.lockutils [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] Releasing lock "refresh_cache-d0310e79-e07a-4581-8cd8-34437cd9ea7d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.977223] env[63372]: DEBUG nova.compute.manager [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Received event network-vif-deleted-5f1c17ad-7301-48b8-8a7d-aacd4717dc74 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 601.977572] env[63372]: DEBUG nova.compute.manager [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Received event network-changed-852395e7-45d6-4d16-8b05-2144329e6dc4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 601.977763] env[63372]: DEBUG nova.compute.manager [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Refreshing instance network info cache due to event network-changed-852395e7-45d6-4d16-8b05-2144329e6dc4. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 601.977978] env[63372]: DEBUG oslo_concurrency.lockutils [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] Acquiring lock "refresh_cache-6a6268e4-0539-45cf-9335-cde31473d1f9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.978133] env[63372]: DEBUG oslo_concurrency.lockutils [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] Acquired lock "refresh_cache-6a6268e4-0539-45cf-9335-cde31473d1f9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.978293] env[63372]: DEBUG nova.network.neutron [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Refreshing network info cache for port 852395e7-45d6-4d16-8b05-2144329e6dc4 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 602.075118] env[63372]: DEBUG nova.network.neutron [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Successfully created port: 5eaf8da0-4d11-4f42-abe3-d5f48f7784cd {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 602.191366] env[63372]: DEBUG nova.scheduler.client.report [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.228570] env[63372]: DEBUG nova.network.neutron [-] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.509832] env[63372]: DEBUG nova.network.neutron [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.685715] env[63372]: DEBUG nova.network.neutron [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.696365] env[63372]: DEBUG nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.703284] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.054s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.703284] env[63372]: ERROR nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7c7784d9-31f5-4c0a-b5d8-93d147d846c0, please check neutron logs for more information. [ 602.703284] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Traceback (most recent call last): [ 602.703284] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.703284] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] self.driver.spawn(context, instance, image_meta, [ 602.703284] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 602.703284] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.703284] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.703284] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] vm_ref = self.build_virtual_machine(instance, [ 602.703529] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.703529] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.703529] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.703529] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] for vif in network_info: [ 602.703529] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.703529] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] return self._sync_wrapper(fn, *args, **kwargs) [ 602.703529] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.703529] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] self.wait() [ 602.703529] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.703529] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] self[:] = self._gt.wait() [ 602.703529] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.703529] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] return self._exit_event.wait() [ 602.703529] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.703837] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] result = hub.switch() [ 602.703837] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.703837] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] return self.greenlet.switch() [ 602.703837] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.703837] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] result = function(*args, **kwargs) [ 602.703837] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.703837] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] return func(*args, **kwargs) [ 602.703837] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.703837] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] raise e [ 602.703837] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.703837] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] nwinfo = self.network_api.allocate_for_instance( [ 602.703837] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.703837] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] created_port_ids = self._update_ports_for_instance( [ 602.704135] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.704135] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] with excutils.save_and_reraise_exception(): [ 602.704135] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.704135] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] self.force_reraise() [ 602.704135] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.704135] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] raise self.value [ 602.704135] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.704135] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] updated_port = self._update_port( [ 602.704135] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.704135] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] _ensure_no_port_binding_failure(port) [ 602.704135] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.704135] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] raise exception.PortBindingFailed(port_id=port['id']) [ 602.704374] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] nova.exception.PortBindingFailed: Binding failed for port 7c7784d9-31f5-4c0a-b5d8-93d147d846c0, please check neutron logs for more information. [ 602.704374] env[63372]: ERROR nova.compute.manager [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] [ 602.704374] env[63372]: DEBUG nova.compute.utils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Binding failed for port 7c7784d9-31f5-4c0a-b5d8-93d147d846c0, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 602.705719] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.874s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.707379] env[63372]: INFO nova.compute.claims [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.710172] env[63372]: DEBUG nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Build of instance 45d6d393-ca23-4e22-bcbf-899ae5037c4f was re-scheduled: Binding failed for port 7c7784d9-31f5-4c0a-b5d8-93d147d846c0, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 602.712074] env[63372]: DEBUG nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 602.712387] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Acquiring lock "refresh_cache-45d6d393-ca23-4e22-bcbf-899ae5037c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.712617] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Acquired lock "refresh_cache-45d6d393-ca23-4e22-bcbf-899ae5037c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.712848] env[63372]: DEBUG nova.network.neutron [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 602.726719] env[63372]: DEBUG nova.virt.hardware [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.727120] env[63372]: DEBUG nova.virt.hardware [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.727120] env[63372]: DEBUG nova.virt.hardware [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.727274] env[63372]: DEBUG nova.virt.hardware [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.727416] env[63372]: DEBUG nova.virt.hardware [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.727560] env[63372]: DEBUG nova.virt.hardware [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.727767] env[63372]: DEBUG nova.virt.hardware [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.728089] env[63372]: DEBUG nova.virt.hardware [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.728163] env[63372]: DEBUG nova.virt.hardware [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.728353] env[63372]: DEBUG nova.virt.hardware [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.728526] env[63372]: DEBUG nova.virt.hardware [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.731798] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8320665b-5eee-4876-b7a9-ad822bb53604 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.738082] env[63372]: INFO nova.compute.manager [-] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Took 1.05 seconds to deallocate network for instance. [ 602.739582] env[63372]: DEBUG nova.compute.claims [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 602.739582] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.747509] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d38842-4e77-428a-b310-d05d9fc59f8f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.772717] env[63372]: DEBUG nova.compute.manager [req-9b0df3b6-d185-490c-9019-6486f1798aaa req-a48892bf-9a06-4b5f-9149-f74a447d0dde service nova] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Received event network-vif-deleted-852395e7-45d6-4d16-8b05-2144329e6dc4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 603.073428] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquiring lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.073672] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.194340] env[63372]: DEBUG oslo_concurrency.lockutils [req-50720147-9248-4f93-8a03-626d21a52c76 req-24aee49a-f5ad-4a71-ac73-6128b40235d5 service nova] Releasing lock "refresh_cache-6a6268e4-0539-45cf-9335-cde31473d1f9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.238932] env[63372]: DEBUG nova.network.neutron [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.325555] env[63372]: DEBUG nova.network.neutron [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.440252] env[63372]: ERROR nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5eaf8da0-4d11-4f42-abe3-d5f48f7784cd, please check neutron logs for more information. [ 603.440252] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 603.440252] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.440252] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 603.440252] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.440252] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 603.440252] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.440252] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 603.440252] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.440252] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 603.440252] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.440252] env[63372]: ERROR nova.compute.manager raise self.value [ 603.440252] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.440252] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 603.440252] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.440252] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 603.440882] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.440882] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 603.440882] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5eaf8da0-4d11-4f42-abe3-d5f48f7784cd, please check neutron logs for more information. [ 603.440882] env[63372]: ERROR nova.compute.manager [ 603.440882] env[63372]: Traceback (most recent call last): [ 603.440882] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 603.440882] env[63372]: listener.cb(fileno) [ 603.440882] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.440882] env[63372]: result = function(*args, **kwargs) [ 603.440882] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.440882] env[63372]: return func(*args, **kwargs) [ 603.440882] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.440882] env[63372]: raise e [ 603.440882] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.440882] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 603.440882] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.440882] env[63372]: created_port_ids = self._update_ports_for_instance( [ 603.440882] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.440882] env[63372]: with excutils.save_and_reraise_exception(): [ 603.440882] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.440882] env[63372]: self.force_reraise() [ 603.440882] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.440882] env[63372]: raise self.value [ 603.440882] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.440882] env[63372]: updated_port = self._update_port( [ 603.440882] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.440882] env[63372]: _ensure_no_port_binding_failure(port) [ 603.440882] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.440882] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 603.441525] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 5eaf8da0-4d11-4f42-abe3-d5f48f7784cd, please check neutron logs for more information. [ 603.441525] env[63372]: Removing descriptor: 19 [ 603.441525] env[63372]: ERROR nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5eaf8da0-4d11-4f42-abe3-d5f48f7784cd, please check neutron logs for more information. [ 603.441525] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Traceback (most recent call last): [ 603.441525] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 603.441525] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] yield resources [ 603.441525] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 603.441525] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] self.driver.spawn(context, instance, image_meta, [ 603.441525] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 603.441525] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] self._vmops.spawn(context, instance, image_meta, injected_files, [ 603.441525] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 603.441525] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] vm_ref = self.build_virtual_machine(instance, [ 603.441874] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 603.441874] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] vif_infos = vmwarevif.get_vif_info(self._session, [ 603.441874] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 603.441874] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] for vif in network_info: [ 603.441874] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 603.441874] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] return self._sync_wrapper(fn, *args, **kwargs) [ 603.441874] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 603.441874] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] self.wait() [ 603.441874] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 603.441874] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] self[:] = self._gt.wait() [ 603.441874] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 603.441874] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] return self._exit_event.wait() [ 603.441874] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 603.442182] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] result = hub.switch() [ 603.442182] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 603.442182] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] return self.greenlet.switch() [ 603.442182] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.442182] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] result = function(*args, **kwargs) [ 603.442182] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.442182] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] return func(*args, **kwargs) [ 603.442182] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.442182] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] raise e [ 603.442182] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.442182] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] nwinfo = self.network_api.allocate_for_instance( [ 603.442182] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.442182] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] created_port_ids = self._update_ports_for_instance( [ 603.442528] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.442528] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] with excutils.save_and_reraise_exception(): [ 603.442528] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.442528] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] self.force_reraise() [ 603.442528] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.442528] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] raise self.value [ 603.442528] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.442528] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] updated_port = self._update_port( [ 603.442528] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.442528] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] _ensure_no_port_binding_failure(port) [ 603.442528] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.442528] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] raise exception.PortBindingFailed(port_id=port['id']) [ 603.447360] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] nova.exception.PortBindingFailed: Binding failed for port 5eaf8da0-4d11-4f42-abe3-d5f48f7784cd, please check neutron logs for more information. [ 603.447360] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] [ 603.447360] env[63372]: INFO nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Terminating instance [ 603.447360] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Acquiring lock "refresh_cache-96c3ce22-9424-4f7f-8810-ee7cfaed0193" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.447360] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Acquired lock "refresh_cache-96c3ce22-9424-4f7f-8810-ee7cfaed0193" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.447360] env[63372]: DEBUG nova.network.neutron [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 603.828789] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Releasing lock "refresh_cache-45d6d393-ca23-4e22-bcbf-899ae5037c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.829015] env[63372]: DEBUG nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 603.829190] env[63372]: DEBUG nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.829362] env[63372]: DEBUG nova.network.neutron [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 603.852198] env[63372]: DEBUG nova.network.neutron [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.962632] env[63372]: DEBUG nova.network.neutron [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.011864] env[63372]: DEBUG nova.network.neutron [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.101276] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-addf8a20-bc37-41ee-984e-ad453cf19fd5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.109393] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d138f3-ae6f-43b6-9ec7-ecb2007483c0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.139756] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ee2fb0-3647-46db-8b63-5d4672508284 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.146754] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc69e086-296e-437f-9e87-739b617a10bc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.159949] env[63372]: DEBUG nova.compute.provider_tree [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.355906] env[63372]: DEBUG nova.network.neutron [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.515228] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Releasing lock "refresh_cache-96c3ce22-9424-4f7f-8810-ee7cfaed0193" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.515785] env[63372]: DEBUG nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 604.515957] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 604.516354] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-650ee1ad-645e-4cf5-b6fe-8ac7af36091c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.525357] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08150952-889f-404d-b11c-f6aedd213b95 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.547978] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 96c3ce22-9424-4f7f-8810-ee7cfaed0193 could not be found. [ 604.548222] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 604.548406] env[63372]: INFO nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Took 0.03 seconds to destroy the instance on the hypervisor. [ 604.549016] env[63372]: DEBUG oslo.service.loopingcall [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.549016] env[63372]: DEBUG nova.compute.manager [-] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.549016] env[63372]: DEBUG nova.network.neutron [-] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 604.563551] env[63372]: DEBUG nova.network.neutron [-] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.663282] env[63372]: DEBUG nova.scheduler.client.report [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.824611] env[63372]: DEBUG nova.compute.manager [req-f3086be7-1369-4bff-b941-e33667c9a810 req-21eb0ca8-36df-4a73-9a79-1f5addec1c09 service nova] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Received event network-changed-5eaf8da0-4d11-4f42-abe3-d5f48f7784cd {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 604.824611] env[63372]: DEBUG nova.compute.manager [req-f3086be7-1369-4bff-b941-e33667c9a810 req-21eb0ca8-36df-4a73-9a79-1f5addec1c09 service nova] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Refreshing instance network info cache due to event network-changed-5eaf8da0-4d11-4f42-abe3-d5f48f7784cd. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 604.824768] env[63372]: DEBUG oslo_concurrency.lockutils [req-f3086be7-1369-4bff-b941-e33667c9a810 req-21eb0ca8-36df-4a73-9a79-1f5addec1c09 service nova] Acquiring lock "refresh_cache-96c3ce22-9424-4f7f-8810-ee7cfaed0193" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.824879] env[63372]: DEBUG oslo_concurrency.lockutils [req-f3086be7-1369-4bff-b941-e33667c9a810 req-21eb0ca8-36df-4a73-9a79-1f5addec1c09 service nova] Acquired lock "refresh_cache-96c3ce22-9424-4f7f-8810-ee7cfaed0193" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.825061] env[63372]: DEBUG nova.network.neutron [req-f3086be7-1369-4bff-b941-e33667c9a810 req-21eb0ca8-36df-4a73-9a79-1f5addec1c09 service nova] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Refreshing network info cache for port 5eaf8da0-4d11-4f42-abe3-d5f48f7784cd {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 604.858365] env[63372]: INFO nova.compute.manager [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] [instance: 45d6d393-ca23-4e22-bcbf-899ae5037c4f] Took 1.03 seconds to deallocate network for instance. [ 605.067187] env[63372]: DEBUG nova.network.neutron [-] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.169059] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.169684] env[63372]: DEBUG nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 605.172299] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.785s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.173772] env[63372]: INFO nova.compute.claims [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 605.347132] env[63372]: DEBUG nova.network.neutron [req-f3086be7-1369-4bff-b941-e33667c9a810 req-21eb0ca8-36df-4a73-9a79-1f5addec1c09 service nova] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.410627] env[63372]: DEBUG nova.network.neutron [req-f3086be7-1369-4bff-b941-e33667c9a810 req-21eb0ca8-36df-4a73-9a79-1f5addec1c09 service nova] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.570067] env[63372]: INFO nova.compute.manager [-] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Took 1.02 seconds to deallocate network for instance. [ 605.572701] env[63372]: DEBUG nova.compute.claims [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 605.572847] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.679499] env[63372]: DEBUG nova.compute.utils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 605.683180] env[63372]: DEBUG nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 605.684091] env[63372]: DEBUG nova.network.neutron [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 605.729580] env[63372]: DEBUG nova.policy [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '365e5ca67d9a4aa3bf794165a97248bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7d32d9e778e4221b6214e079d5242ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 605.904471] env[63372]: INFO nova.scheduler.client.report [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Deleted allocations for instance 45d6d393-ca23-4e22-bcbf-899ae5037c4f [ 605.915828] env[63372]: DEBUG oslo_concurrency.lockutils [req-f3086be7-1369-4bff-b941-e33667c9a810 req-21eb0ca8-36df-4a73-9a79-1f5addec1c09 service nova] Releasing lock "refresh_cache-96c3ce22-9424-4f7f-8810-ee7cfaed0193" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.915828] env[63372]: DEBUG nova.compute.manager [req-f3086be7-1369-4bff-b941-e33667c9a810 req-21eb0ca8-36df-4a73-9a79-1f5addec1c09 service nova] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Received event network-vif-deleted-5eaf8da0-4d11-4f42-abe3-d5f48f7784cd {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 606.011702] env[63372]: DEBUG nova.network.neutron [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Successfully created port: 40531f78-08e2-4c40-bfa7-c896e07e4538 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.184455] env[63372]: DEBUG nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 606.414890] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8aaf117b-7b48-4195-bd87-fe05b7589447 tempest-DeleteServersAdminTestJSON-1054341311 tempest-DeleteServersAdminTestJSON-1054341311-project-member] Lock "45d6d393-ca23-4e22-bcbf-899ae5037c4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.404s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.571786] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2377775-7a96-4ce3-9de9-9b6c2916df2b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.579858] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956606a1-3c30-453f-871a-ea7a04f0372a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.611460] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f8a946-59ba-4f7d-be62-72692be4081b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.618928] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1858705a-3c7b-4046-b671-23947440935e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.635392] env[63372]: DEBUG nova.compute.provider_tree [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.917502] env[63372]: DEBUG nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 607.135079] env[63372]: DEBUG nova.compute.manager [req-07c2d468-800d-4efc-9b53-76286fefde27 req-8a135387-871d-4d16-a2e0-101dfbbb1a44 service nova] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Received event network-changed-40531f78-08e2-4c40-bfa7-c896e07e4538 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 607.139030] env[63372]: DEBUG nova.compute.manager [req-07c2d468-800d-4efc-9b53-76286fefde27 req-8a135387-871d-4d16-a2e0-101dfbbb1a44 service nova] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Refreshing instance network info cache due to event network-changed-40531f78-08e2-4c40-bfa7-c896e07e4538. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 607.139030] env[63372]: DEBUG oslo_concurrency.lockutils [req-07c2d468-800d-4efc-9b53-76286fefde27 req-8a135387-871d-4d16-a2e0-101dfbbb1a44 service nova] Acquiring lock "refresh_cache-c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.139030] env[63372]: DEBUG oslo_concurrency.lockutils [req-07c2d468-800d-4efc-9b53-76286fefde27 req-8a135387-871d-4d16-a2e0-101dfbbb1a44 service nova] Acquired lock "refresh_cache-c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.139030] env[63372]: DEBUG nova.network.neutron [req-07c2d468-800d-4efc-9b53-76286fefde27 req-8a135387-871d-4d16-a2e0-101dfbbb1a44 service nova] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Refreshing network info cache for port 40531f78-08e2-4c40-bfa7-c896e07e4538 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 607.139522] env[63372]: DEBUG nova.scheduler.client.report [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 607.196778] env[63372]: DEBUG nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 607.240072] env[63372]: DEBUG nova.virt.hardware [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 607.241032] env[63372]: DEBUG nova.virt.hardware [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 607.241032] env[63372]: DEBUG nova.virt.hardware [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 607.241032] env[63372]: DEBUG nova.virt.hardware [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 607.241032] env[63372]: DEBUG nova.virt.hardware [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 607.241317] env[63372]: DEBUG nova.virt.hardware [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 607.241389] env[63372]: DEBUG nova.virt.hardware [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 607.241537] env[63372]: DEBUG nova.virt.hardware [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 607.241692] env[63372]: DEBUG nova.virt.hardware [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 607.241842] env[63372]: DEBUG nova.virt.hardware [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 607.242009] env[63372]: DEBUG nova.virt.hardware [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 607.242874] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf1c2314-f660-4e1c-ab87-6d95f39dfc14 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.251358] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15c02ae-a4e2-4256-8309-9be89063cd1e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.332907] env[63372]: ERROR nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 40531f78-08e2-4c40-bfa7-c896e07e4538, please check neutron logs for more information. [ 607.332907] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 607.332907] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.332907] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 607.332907] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.332907] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 607.332907] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.332907] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 607.332907] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.332907] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 607.332907] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.332907] env[63372]: ERROR nova.compute.manager raise self.value [ 607.332907] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.332907] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 607.332907] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.332907] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 607.333301] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.333301] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 607.333301] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 40531f78-08e2-4c40-bfa7-c896e07e4538, please check neutron logs for more information. [ 607.333301] env[63372]: ERROR nova.compute.manager [ 607.333301] env[63372]: Traceback (most recent call last): [ 607.333301] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 607.333301] env[63372]: listener.cb(fileno) [ 607.333301] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.333301] env[63372]: result = function(*args, **kwargs) [ 607.333301] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.333301] env[63372]: return func(*args, **kwargs) [ 607.333301] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.333301] env[63372]: raise e [ 607.333301] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.333301] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 607.333301] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.333301] env[63372]: created_port_ids = self._update_ports_for_instance( [ 607.333301] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.333301] env[63372]: with excutils.save_and_reraise_exception(): [ 607.333301] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.333301] env[63372]: self.force_reraise() [ 607.333301] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.333301] env[63372]: raise self.value [ 607.333301] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.333301] env[63372]: updated_port = self._update_port( [ 607.333301] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.333301] env[63372]: _ensure_no_port_binding_failure(port) [ 607.333301] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.333301] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 607.333897] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 40531f78-08e2-4c40-bfa7-c896e07e4538, please check neutron logs for more information. [ 607.333897] env[63372]: Removing descriptor: 19 [ 607.336056] env[63372]: ERROR nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 40531f78-08e2-4c40-bfa7-c896e07e4538, please check neutron logs for more information. [ 607.336056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Traceback (most recent call last): [ 607.336056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 607.336056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] yield resources [ 607.336056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 607.336056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] self.driver.spawn(context, instance, image_meta, [ 607.336056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 607.336056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.336056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.336056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] vm_ref = self.build_virtual_machine(instance, [ 607.336056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.336494] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.336494] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.336494] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] for vif in network_info: [ 607.336494] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.336494] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] return self._sync_wrapper(fn, *args, **kwargs) [ 607.336494] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.336494] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] self.wait() [ 607.336494] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.336494] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] self[:] = self._gt.wait() [ 607.336494] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.336494] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] return self._exit_event.wait() [ 607.336494] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.336494] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] result = hub.switch() [ 607.336780] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.336780] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] return self.greenlet.switch() [ 607.336780] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.336780] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] result = function(*args, **kwargs) [ 607.336780] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.336780] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] return func(*args, **kwargs) [ 607.336780] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.336780] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] raise e [ 607.336780] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.336780] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] nwinfo = self.network_api.allocate_for_instance( [ 607.336780] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.336780] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] created_port_ids = self._update_ports_for_instance( [ 607.336780] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.337056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] with excutils.save_and_reraise_exception(): [ 607.337056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.337056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] self.force_reraise() [ 607.337056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.337056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] raise self.value [ 607.337056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.337056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] updated_port = self._update_port( [ 607.337056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.337056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] _ensure_no_port_binding_failure(port) [ 607.337056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.337056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] raise exception.PortBindingFailed(port_id=port['id']) [ 607.337056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] nova.exception.PortBindingFailed: Binding failed for port 40531f78-08e2-4c40-bfa7-c896e07e4538, please check neutron logs for more information. [ 607.337056] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] [ 607.337392] env[63372]: INFO nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Terminating instance [ 607.340076] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Acquiring lock "refresh_cache-c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.441303] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.646248] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.646248] env[63372]: DEBUG nova.compute.manager [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 607.648702] env[63372]: DEBUG oslo_concurrency.lockutils [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.131s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.650060] env[63372]: INFO nova.compute.claims [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 607.667543] env[63372]: DEBUG nova.network.neutron [req-07c2d468-800d-4efc-9b53-76286fefde27 req-8a135387-871d-4d16-a2e0-101dfbbb1a44 service nova] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.746364] env[63372]: DEBUG nova.network.neutron [req-07c2d468-800d-4efc-9b53-76286fefde27 req-8a135387-871d-4d16-a2e0-101dfbbb1a44 service nova] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.160566] env[63372]: DEBUG nova.compute.utils [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 608.164257] env[63372]: DEBUG nova.compute.manager [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Not allocating networking since 'none' was specified. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 608.249297] env[63372]: DEBUG oslo_concurrency.lockutils [req-07c2d468-800d-4efc-9b53-76286fefde27 req-8a135387-871d-4d16-a2e0-101dfbbb1a44 service nova] Releasing lock "refresh_cache-c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.249705] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Acquired lock "refresh_cache-c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.249887] env[63372]: DEBUG nova.network.neutron [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 608.665031] env[63372]: DEBUG nova.compute.manager [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 608.781636] env[63372]: DEBUG nova.network.neutron [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.023167] env[63372]: DEBUG nova.network.neutron [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.115575] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f846a888-5fe8-4d5d-8924-8c0937904466 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.124343] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00dca804-5c9c-47d2-950a-2b62692533fa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.161648] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb95807-638b-4c31-b6d8-e10def3d448a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.169920] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4586c8-a54b-4f75-b3ee-536024e6779f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.187382] env[63372]: DEBUG nova.compute.provider_tree [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.196593] env[63372]: DEBUG nova.compute.manager [req-3fb867db-0e92-4561-a979-b511cb88612f req-0bcbf0a2-ca9d-49c6-8054-cb20c5decba6 service nova] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Received event network-vif-deleted-40531f78-08e2-4c40-bfa7-c896e07e4538 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 609.526308] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Releasing lock "refresh_cache-c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.526592] env[63372]: DEBUG nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 609.526795] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 609.527278] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-78df6194-70e0-4ba9-a553-10e69d96cea1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.541776] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7afe00c5-20c4-48ca-92e6-3704aee2adbd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.570129] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08 could not be found. [ 609.570602] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 609.570886] env[63372]: INFO nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Took 0.04 seconds to destroy the instance on the hypervisor. [ 609.571265] env[63372]: DEBUG oslo.service.loopingcall [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 609.571586] env[63372]: DEBUG nova.compute.manager [-] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 609.571761] env[63372]: DEBUG nova.network.neutron [-] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 609.593730] env[63372]: DEBUG nova.network.neutron [-] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.679874] env[63372]: DEBUG nova.compute.manager [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 609.696653] env[63372]: DEBUG nova.scheduler.client.report [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.718019] env[63372]: DEBUG nova.virt.hardware [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 609.718019] env[63372]: DEBUG nova.virt.hardware [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 609.718019] env[63372]: DEBUG nova.virt.hardware [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 609.718233] env[63372]: DEBUG nova.virt.hardware [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 609.718233] env[63372]: DEBUG nova.virt.hardware [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 609.718233] env[63372]: DEBUG nova.virt.hardware [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 609.718233] env[63372]: DEBUG nova.virt.hardware [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 609.718233] env[63372]: DEBUG nova.virt.hardware [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 609.718368] env[63372]: DEBUG nova.virt.hardware [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 609.718368] env[63372]: DEBUG nova.virt.hardware [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 609.718368] env[63372]: DEBUG nova.virt.hardware [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 609.719069] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da58a8c-2dbc-4478-81b2-3d26b098198e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.729033] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b6ba859-48b6-4164-8aab-a79cd2a6b748 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.744573] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Instance VIF info [] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 609.754776] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 609.755670] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b6adfaab-3532-4c8a-9b8c-49a1c66f9d7e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.767353] env[63372]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 609.768179] env[63372]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63372) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 609.768179] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 609.768359] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Creating folder: Project (5e54b0f91ebf4f3495d619f58e19b8a1). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 609.768441] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4223c90a-21ef-4048-989c-b0e02e2ec96f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.777776] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Created folder: Project (5e54b0f91ebf4f3495d619f58e19b8a1) in parent group-v227230. [ 609.778831] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Creating folder: Instances. Parent ref: group-v227235. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 609.778831] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ca52695a-1627-4ee7-bc83-81a86d411f59 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.787921] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Created folder: Instances in parent group-v227235. [ 609.787921] env[63372]: DEBUG oslo.service.loopingcall [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 609.788127] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8986e782-4d63-4491-8100-4341a3149812] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 609.788188] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-33c62f93-3599-4cf3-b293-355021f7213a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.809096] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 609.809096] env[63372]: value = "task-1023647" [ 609.809096] env[63372]: _type = "Task" [ 609.809096] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.818625] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023647, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.097742] env[63372]: DEBUG nova.network.neutron [-] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.201576] env[63372]: DEBUG oslo_concurrency.lockutils [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.552s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.201576] env[63372]: DEBUG nova.compute.manager [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 610.206232] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.763s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.209584] env[63372]: INFO nova.compute.claims [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 610.320598] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023647, 'name': CreateVM_Task, 'duration_secs': 0.266019} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.321214] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8986e782-4d63-4491-8100-4341a3149812] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 610.322455] env[63372]: DEBUG oslo_vmware.service [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff34123-f5ba-4379-bd72-6787a8f57627 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.328571] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.328756] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.329458] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 610.329700] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a83500a-0f75-4ab8-b476-5e145c49f622 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.334547] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for the task: (returnval){ [ 610.334547] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52643e20-2456-b117-6233-0874ad44e1f9" [ 610.334547] env[63372]: _type = "Task" [ 610.334547] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.342556] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52643e20-2456-b117-6233-0874ad44e1f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.553592] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Acquiring lock "5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.553868] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Lock "5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.600777] env[63372]: INFO nova.compute.manager [-] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Took 1.03 seconds to deallocate network for instance. [ 610.603448] env[63372]: DEBUG nova.compute.claims [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 610.603704] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.713815] env[63372]: DEBUG nova.compute.utils [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 610.716198] env[63372]: DEBUG nova.compute.manager [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Not allocating networking since 'none' was specified. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 610.845965] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.847250] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 610.847250] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.847250] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.847250] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 610.847512] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b962322-fbd6-4a4d-ad0f-1ea7b3adb048 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.863704] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 610.863883] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 610.864669] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e643fc80-9009-41e7-ab2b-07ed723f3a1e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.870820] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd3074b9-190f-479e-8f44-8ed4462b87cc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.875911] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for the task: (returnval){ [ 610.875911] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52114071-28b5-0132-930d-5c055fc2e7be" [ 610.875911] env[63372]: _type = "Task" [ 610.875911] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.883321] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52114071-28b5-0132-930d-5c055fc2e7be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.222020] env[63372]: DEBUG nova.compute.manager [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 611.393982] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Preparing fetch location {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 611.394270] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Creating directory with path [datastore2] vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6/d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 611.394747] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-38d644d6-2c12-47a8-8117-5a524e48405c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.416497] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Created directory with path [datastore2] vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6/d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 611.417048] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Fetch image to [datastore2] vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 611.417048] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Downloading image file data d7ae1717-77b2-47f6-9acd-b27bac221f7c to [datastore2] vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk on the data store datastore2 {{(pid=63372) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 611.419413] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581fae50-5863-43ad-9e3d-78dd4e0d30b8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.431150] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7706914-c03e-4b1c-b050-158ffc64c38d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.443223] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b513c4ef-9572-4e18-8f78-9d48ec365fd0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.479388] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd08eb0-34e2-48ca-b52d-4554150eed1a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.485973] env[63372]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-768f80d6-46f2-4cca-8992-319531849026 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.514416] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Downloading image file data d7ae1717-77b2-47f6-9acd-b27bac221f7c to the data store datastore2 {{(pid=63372) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 611.575073] env[63372]: DEBUG oslo_vmware.rw_handles [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63372) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 611.748405] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcca48ea-660e-4f81-9a62-ec57aed99426 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.761083] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83547716-9ba5-4e04-ae9a-5662d43fa8a2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.798351] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7af4cae-0cdf-4106-acfd-5f5681da7207 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.807366] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e6cfdd-371f-4bbb-81a9-241ca1073a64 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.826136] env[63372]: DEBUG nova.compute.provider_tree [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.237169] env[63372]: DEBUG nova.compute.manager [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 612.267879] env[63372]: DEBUG nova.virt.hardware [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 612.268126] env[63372]: DEBUG nova.virt.hardware [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 612.268289] env[63372]: DEBUG nova.virt.hardware [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 612.268469] env[63372]: DEBUG nova.virt.hardware [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 612.268609] env[63372]: DEBUG nova.virt.hardware [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 612.268752] env[63372]: DEBUG nova.virt.hardware [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 612.269031] env[63372]: DEBUG nova.virt.hardware [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 612.269225] env[63372]: DEBUG nova.virt.hardware [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 612.269401] env[63372]: DEBUG nova.virt.hardware [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 612.269564] env[63372]: DEBUG nova.virt.hardware [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 612.269770] env[63372]: DEBUG nova.virt.hardware [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 612.270786] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce843a0-3627-45ab-a5d4-93fb5fd5efc0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.276737] env[63372]: DEBUG oslo_vmware.rw_handles [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Completed reading data from the image iterator. {{(pid=63372) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 612.276737] env[63372]: DEBUG oslo_vmware.rw_handles [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 612.280960] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d4de62-c8b0-49a5-a93a-221fb58fb6e2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.296385] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Instance VIF info [] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 612.301994] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Creating folder: Project (b529b17076e844af81b2702637e0bc68). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 612.302284] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1154f03-ed98-45d7-a1ed-964674144e19 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.314035] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Created folder: Project (b529b17076e844af81b2702637e0bc68) in parent group-v227230. [ 612.314035] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Creating folder: Instances. Parent ref: group-v227238. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 612.314035] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-01e13325-aca6-420a-8e4a-21d9d61a6d47 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.321793] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Created folder: Instances in parent group-v227238. [ 612.322060] env[63372]: DEBUG oslo.service.loopingcall [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.322259] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 612.322464] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-01fa7c11-5823-4792-b80f-7e4b319acace {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.335184] env[63372]: DEBUG nova.scheduler.client.report [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.345819] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 612.345819] env[63372]: value = "task-1023650" [ 612.345819] env[63372]: _type = "Task" [ 612.345819] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.359060] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023650, 'name': CreateVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.424602] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Downloaded image file data d7ae1717-77b2-47f6-9acd-b27bac221f7c to vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk on the data store datastore2 {{(pid=63372) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 612.426358] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Caching image {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 612.426609] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Copying Virtual Disk [datastore2] vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk to [datastore2] vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 612.427334] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e172edbf-f0b8-421e-b815-d8a0df08d1aa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.435054] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for the task: (returnval){ [ 612.435054] env[63372]: value = "task-1023651" [ 612.435054] env[63372]: _type = "Task" [ 612.435054] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.443978] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023651, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.843134] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.634s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.843134] env[63372]: DEBUG nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 612.844603] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.436s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.846594] env[63372]: INFO nova.compute.claims [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.864424] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023650, 'name': CreateVM_Task, 'duration_secs': 0.272754} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.864606] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 612.865777] env[63372]: DEBUG oslo_concurrency.lockutils [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.866044] env[63372]: DEBUG oslo_concurrency.lockutils [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.866456] env[63372]: DEBUG oslo_concurrency.lockutils [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 612.867664] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e7779c5-e305-4300-a23d-c959ead72aac {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.877576] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 612.877576] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52d9ab2a-be29-b80b-3637-4741374a0301" [ 612.877576] env[63372]: _type = "Task" [ 612.877576] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.890943] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52d9ab2a-be29-b80b-3637-4741374a0301, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.947257] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023651, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.351404] env[63372]: DEBUG nova.compute.utils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 613.352816] env[63372]: DEBUG nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 613.356182] env[63372]: DEBUG nova.network.neutron [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 613.390981] env[63372]: DEBUG oslo_concurrency.lockutils [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.391280] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 613.391501] env[63372]: DEBUG oslo_concurrency.lockutils [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.413539] env[63372]: DEBUG nova.policy [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f8e364e3aed40248cc0d11741f76154', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b4548b52cd704cc0b054e2f7d1562b9a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 613.452142] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023651, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.678495} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.452425] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Copied Virtual Disk [datastore2] vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk to [datastore2] vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 613.452614] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Deleting the datastore file [datastore2] vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 613.452872] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6e4673b1-686c-43da-b3eb-174fd12579f5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.460380] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for the task: (returnval){ [ 613.460380] env[63372]: value = "task-1023652" [ 613.460380] env[63372]: _type = "Task" [ 613.460380] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.470706] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023652, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.741104] env[63372]: DEBUG nova.network.neutron [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Successfully created port: fb497c35-8c18-42a5-81e0-b59b655122c5 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 613.857745] env[63372]: DEBUG nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 613.977525] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023652, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023205} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.977692] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 613.977893] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Moving file from [datastore2] vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6/d7ae1717-77b2-47f6-9acd-b27bac221f7c to [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c. {{(pid=63372) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 613.978323] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-ea235206-b239-4d0c-9ce8-c5b654e7c11f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.987452] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for the task: (returnval){ [ 613.987452] env[63372]: value = "task-1023653" [ 613.987452] env[63372]: _type = "Task" [ 613.987452] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.996693] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023653, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.156869] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "63b9095c-fb49-4dc1-a6dc-96529aeaab81" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.157112] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "63b9095c-fb49-4dc1-a6dc-96529aeaab81" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.304484] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3043ac0f-51f2-459f-862a-19b0a66bd5bc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.314017] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76078ffb-e50e-4d3c-9e8f-69d8d4764b4b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.347394] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5784fe54-05c3-479d-80c1-0fad5d501672 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.355970] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0323510f-3b4e-4940-a6e3-75679cff77ae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.372709] env[63372]: DEBUG nova.compute.provider_tree [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.464134] env[63372]: DEBUG nova.compute.manager [req-680bfa21-b5db-499e-a1fd-bb9555b28da7 req-a3696553-8661-4c63-aeb5-3110674c1d3c service nova] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Received event network-changed-fb497c35-8c18-42a5-81e0-b59b655122c5 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 614.464134] env[63372]: DEBUG nova.compute.manager [req-680bfa21-b5db-499e-a1fd-bb9555b28da7 req-a3696553-8661-4c63-aeb5-3110674c1d3c service nova] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Refreshing instance network info cache due to event network-changed-fb497c35-8c18-42a5-81e0-b59b655122c5. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 614.464564] env[63372]: DEBUG oslo_concurrency.lockutils [req-680bfa21-b5db-499e-a1fd-bb9555b28da7 req-a3696553-8661-4c63-aeb5-3110674c1d3c service nova] Acquiring lock "refresh_cache-bd8cf6ba-8039-4c02-bda2-13c7d14f8903" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.464862] env[63372]: DEBUG oslo_concurrency.lockutils [req-680bfa21-b5db-499e-a1fd-bb9555b28da7 req-a3696553-8661-4c63-aeb5-3110674c1d3c service nova] Acquired lock "refresh_cache-bd8cf6ba-8039-4c02-bda2-13c7d14f8903" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.465239] env[63372]: DEBUG nova.network.neutron [req-680bfa21-b5db-499e-a1fd-bb9555b28da7 req-a3696553-8661-4c63-aeb5-3110674c1d3c service nova] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Refreshing network info cache for port fb497c35-8c18-42a5-81e0-b59b655122c5 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 614.498935] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023653, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.023857} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.499330] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] File moved {{(pid=63372) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 614.499600] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Cleaning up location [datastore2] vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6 {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 614.499843] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Deleting the datastore file [datastore2] vmware_temp/9d1483e8-1835-4f17-8919-a3820cf5fec6 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 614.500423] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e1d77540-0019-4171-8f22-e713f24e97d3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.506973] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for the task: (returnval){ [ 614.506973] env[63372]: value = "task-1023654" [ 614.506973] env[63372]: _type = "Task" [ 614.506973] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.516703] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023654, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.662246] env[63372]: ERROR nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fb497c35-8c18-42a5-81e0-b59b655122c5, please check neutron logs for more information. [ 614.662246] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 614.662246] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.662246] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 614.662246] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.662246] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 614.662246] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.662246] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 614.662246] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.662246] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 614.662246] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.662246] env[63372]: ERROR nova.compute.manager raise self.value [ 614.662246] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.662246] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 614.662246] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.662246] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 614.662865] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.662865] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 614.662865] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fb497c35-8c18-42a5-81e0-b59b655122c5, please check neutron logs for more information. [ 614.662865] env[63372]: ERROR nova.compute.manager [ 614.662865] env[63372]: Traceback (most recent call last): [ 614.662865] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 614.662865] env[63372]: listener.cb(fileno) [ 614.662865] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.662865] env[63372]: result = function(*args, **kwargs) [ 614.662865] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.662865] env[63372]: return func(*args, **kwargs) [ 614.662865] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.662865] env[63372]: raise e [ 614.662865] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.662865] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 614.662865] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.662865] env[63372]: created_port_ids = self._update_ports_for_instance( [ 614.662865] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.662865] env[63372]: with excutils.save_and_reraise_exception(): [ 614.662865] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.662865] env[63372]: self.force_reraise() [ 614.662865] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.662865] env[63372]: raise self.value [ 614.662865] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.662865] env[63372]: updated_port = self._update_port( [ 614.662865] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.662865] env[63372]: _ensure_no_port_binding_failure(port) [ 614.662865] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.662865] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 614.663953] env[63372]: nova.exception.PortBindingFailed: Binding failed for port fb497c35-8c18-42a5-81e0-b59b655122c5, please check neutron logs for more information. [ 614.663953] env[63372]: Removing descriptor: 19 [ 614.875652] env[63372]: DEBUG nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 614.878253] env[63372]: DEBUG nova.scheduler.client.report [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.912778] env[63372]: DEBUG nova.virt.hardware [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 614.912778] env[63372]: DEBUG nova.virt.hardware [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 614.912778] env[63372]: DEBUG nova.virt.hardware [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 614.912778] env[63372]: DEBUG nova.virt.hardware [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 614.913192] env[63372]: DEBUG nova.virt.hardware [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 614.913192] env[63372]: DEBUG nova.virt.hardware [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 614.913192] env[63372]: DEBUG nova.virt.hardware [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 614.913192] env[63372]: DEBUG nova.virt.hardware [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 614.913192] env[63372]: DEBUG nova.virt.hardware [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 614.913350] env[63372]: DEBUG nova.virt.hardware [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 614.913350] env[63372]: DEBUG nova.virt.hardware [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 614.914131] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c12eaf9-ed3d-4807-b9fb-8d8b34ae2bcf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.923890] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01247766-dd55-42b3-a324-ad793ec132af {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.941209] env[63372]: ERROR nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fb497c35-8c18-42a5-81e0-b59b655122c5, please check neutron logs for more information. [ 614.941209] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Traceback (most recent call last): [ 614.941209] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 614.941209] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] yield resources [ 614.941209] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.941209] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] self.driver.spawn(context, instance, image_meta, [ 614.941209] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 614.941209] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.941209] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.941209] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] vm_ref = self.build_virtual_machine(instance, [ 614.941209] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.941575] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.941575] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.941575] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] for vif in network_info: [ 614.941575] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.941575] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] return self._sync_wrapper(fn, *args, **kwargs) [ 614.941575] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.941575] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] self.wait() [ 614.941575] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.941575] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] self[:] = self._gt.wait() [ 614.941575] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.941575] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] return self._exit_event.wait() [ 614.941575] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 614.941575] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] current.throw(*self._exc) [ 614.941884] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.941884] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] result = function(*args, **kwargs) [ 614.941884] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.941884] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] return func(*args, **kwargs) [ 614.941884] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.941884] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] raise e [ 614.941884] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.941884] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] nwinfo = self.network_api.allocate_for_instance( [ 614.941884] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.941884] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] created_port_ids = self._update_ports_for_instance( [ 614.941884] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.941884] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] with excutils.save_and_reraise_exception(): [ 614.941884] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.942197] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] self.force_reraise() [ 614.942197] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.942197] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] raise self.value [ 614.942197] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.942197] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] updated_port = self._update_port( [ 614.942197] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.942197] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] _ensure_no_port_binding_failure(port) [ 614.942197] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.942197] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] raise exception.PortBindingFailed(port_id=port['id']) [ 614.942197] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] nova.exception.PortBindingFailed: Binding failed for port fb497c35-8c18-42a5-81e0-b59b655122c5, please check neutron logs for more information. [ 614.942197] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] [ 614.942197] env[63372]: INFO nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Terminating instance [ 614.946344] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "refresh_cache-bd8cf6ba-8039-4c02-bda2-13c7d14f8903" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.989658] env[63372]: DEBUG nova.network.neutron [req-680bfa21-b5db-499e-a1fd-bb9555b28da7 req-a3696553-8661-4c63-aeb5-3110674c1d3c service nova] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.018442] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023654, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024781} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.018702] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 615.019443] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-313c99f7-30ff-4b84-a824-16d010680a5c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.024700] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for the task: (returnval){ [ 615.024700] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52123d2f-5cf2-f396-bc7c-776d7ccb0f71" [ 615.024700] env[63372]: _type = "Task" [ 615.024700] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.035707] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52123d2f-5cf2-f396-bc7c-776d7ccb0f71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.075676] env[63372]: DEBUG nova.network.neutron [req-680bfa21-b5db-499e-a1fd-bb9555b28da7 req-a3696553-8661-4c63-aeb5-3110674c1d3c service nova] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.383978] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.539s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.384585] env[63372]: DEBUG nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 615.387216] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.694s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.535558] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52123d2f-5cf2-f396-bc7c-776d7ccb0f71, 'name': SearchDatastore_Task, 'duration_secs': 0.009719} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.535801] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.536071] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 8986e782-4d63-4491-8100-4341a3149812/8986e782-4d63-4491-8100-4341a3149812.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 615.536339] env[63372]: DEBUG oslo_concurrency.lockutils [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.536523] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 615.536722] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-96f8e214-7c74-4d64-9a4d-77247bdc7d9a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.538589] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-71532f9e-71c2-44c2-a551-ad03a7bd787a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.545575] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for the task: (returnval){ [ 615.545575] env[63372]: value = "task-1023655" [ 615.545575] env[63372]: _type = "Task" [ 615.545575] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.548995] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 615.549214] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 615.549960] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fec39bfd-9cff-4dd6-bae4-f9122591ebde {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.559825] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 615.559825] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e84a7f-3411-0749-16c4-221d885e39ab" [ 615.559825] env[63372]: _type = "Task" [ 615.559825] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.559825] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023655, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.566232] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e84a7f-3411-0749-16c4-221d885e39ab, 'name': SearchDatastore_Task, 'duration_secs': 0.007694} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.566965] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab8d1ff1-a7c0-4f48-9b6e-ef6a5cc3cb11 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.571518] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 615.571518] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521e346f-ba48-98d4-733e-95bc2101f707" [ 615.571518] env[63372]: _type = "Task" [ 615.571518] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.580510] env[63372]: DEBUG oslo_concurrency.lockutils [req-680bfa21-b5db-499e-a1fd-bb9555b28da7 req-a3696553-8661-4c63-aeb5-3110674c1d3c service nova] Releasing lock "refresh_cache-bd8cf6ba-8039-4c02-bda2-13c7d14f8903" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.580867] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521e346f-ba48-98d4-733e-95bc2101f707, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.581117] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired lock "refresh_cache-bd8cf6ba-8039-4c02-bda2-13c7d14f8903" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.581312] env[63372]: DEBUG nova.network.neutron [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 615.891918] env[63372]: DEBUG nova.compute.utils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.897528] env[63372]: DEBUG nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 615.897868] env[63372]: DEBUG nova.network.neutron [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 615.940829] env[63372]: DEBUG nova.policy [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f49aa6c08444402d92a8094a4e262b79', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '061189d394c040c2b63ac28609710bc1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 616.057217] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023655, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459359} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.059718] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 8986e782-4d63-4491-8100-4341a3149812/8986e782-4d63-4491-8100-4341a3149812.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 616.059932] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 616.060367] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-48c68e76-66c6-48f4-87b0-03f67ed6372f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.067167] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for the task: (returnval){ [ 616.067167] env[63372]: value = "task-1023656" [ 616.067167] env[63372]: _type = "Task" [ 616.067167] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.078328] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023656, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.085459] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521e346f-ba48-98d4-733e-95bc2101f707, 'name': SearchDatastore_Task, 'duration_secs': 0.00724} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.088014] env[63372]: DEBUG oslo_concurrency.lockutils [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.088313] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 715408b8-0e76-4b61-a342-b168377cb288/715408b8-0e76-4b61-a342-b168377cb288.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 616.089354] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-468e51d6-a6a1-4128-be0c-5a18ff7655bf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.095491] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 616.095491] env[63372]: value = "task-1023657" [ 616.095491] env[63372]: _type = "Task" [ 616.095491] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.109074] env[63372]: DEBUG nova.network.neutron [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.112232] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023657, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.245161] env[63372]: DEBUG nova.network.neutron [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.302113] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98644e16-fb7f-48be-9c04-2faa3297adf3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.304798] env[63372]: DEBUG nova.network.neutron [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Successfully created port: 98f102f8-7a5b-4648-b3ce-a66979a361b5 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 616.312606] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e54dc818-72f0-41ac-84fe-7edded795968 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.346813] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f7a1b6-33b6-44c5-9684-a6a3859aa300 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.359030] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b270db-adf4-4195-8a8a-0fc1433562e8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.974099] env[63372]: DEBUG nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 616.978333] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Releasing lock "refresh_cache-bd8cf6ba-8039-4c02-bda2-13c7d14f8903" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.978698] env[63372]: DEBUG nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 616.978877] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 616.979349] env[63372]: DEBUG nova.compute.provider_tree [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.983707] env[63372]: DEBUG nova.compute.manager [req-1eb25f03-0b36-4c36-99a5-782e1b855bd5 req-3a49a052-2755-42f9-ac4c-d66c6294d5c4 service nova] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Received event network-vif-deleted-fb497c35-8c18-42a5-81e0-b59b655122c5 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 616.987599] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ff218b8-34bc-4d39-a6cd-a62310ff8691 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.990796] env[63372]: DEBUG nova.scheduler.client.report [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 617.002486] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023656, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065366} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.008022] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 617.008022] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023657, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444238} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.008022] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e8df19-e015-4af0-ae64-53114b0f41c4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.009538] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 715408b8-0e76-4b61-a342-b168377cb288/715408b8-0e76-4b61-a342-b168377cb288.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 617.009731] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 617.011526] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-22c46cad-cd6f-4890-a960-0756a719f785 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.016808] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d62885f-adb7-429a-aa96-7367d2e432a0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.047203] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Reconfiguring VM instance instance-00000016 to attach disk [datastore2] 8986e782-4d63-4491-8100-4341a3149812/8986e782-4d63-4491-8100-4341a3149812.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 617.048417] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6fb5eea-0c24-402a-b4e9-5f1323605e4b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.062593] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 617.062593] env[63372]: value = "task-1023658" [ 617.062593] env[63372]: _type = "Task" [ 617.062593] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.067921] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bd8cf6ba-8039-4c02-bda2-13c7d14f8903 could not be found. [ 617.068145] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 617.068322] env[63372]: INFO nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Took 0.09 seconds to destroy the instance on the hypervisor. [ 617.068554] env[63372]: DEBUG oslo.service.loopingcall [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 617.072394] env[63372]: DEBUG nova.compute.manager [-] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 617.072503] env[63372]: DEBUG nova.network.neutron [-] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 617.075579] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for the task: (returnval){ [ 617.075579] env[63372]: value = "task-1023659" [ 617.075579] env[63372]: _type = "Task" [ 617.075579] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.082167] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023658, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.086899] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023659, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.093705] env[63372]: DEBUG nova.network.neutron [-] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.320479] env[63372]: ERROR nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 98f102f8-7a5b-4648-b3ce-a66979a361b5, please check neutron logs for more information. [ 617.320479] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 617.320479] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.320479] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 617.320479] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.320479] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 617.320479] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.320479] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 617.320479] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.320479] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 617.320479] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.320479] env[63372]: ERROR nova.compute.manager raise self.value [ 617.320479] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.320479] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 617.320479] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.320479] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 617.321029] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.321029] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 617.321029] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 98f102f8-7a5b-4648-b3ce-a66979a361b5, please check neutron logs for more information. [ 617.321029] env[63372]: ERROR nova.compute.manager [ 617.321029] env[63372]: Traceback (most recent call last): [ 617.321029] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 617.321029] env[63372]: listener.cb(fileno) [ 617.321029] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.321029] env[63372]: result = function(*args, **kwargs) [ 617.321029] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.321029] env[63372]: return func(*args, **kwargs) [ 617.321029] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.321029] env[63372]: raise e [ 617.321029] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.321029] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 617.321029] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.321029] env[63372]: created_port_ids = self._update_ports_for_instance( [ 617.321029] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.321029] env[63372]: with excutils.save_and_reraise_exception(): [ 617.321029] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.321029] env[63372]: self.force_reraise() [ 617.321029] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.321029] env[63372]: raise self.value [ 617.321029] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.321029] env[63372]: updated_port = self._update_port( [ 617.321029] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.321029] env[63372]: _ensure_no_port_binding_failure(port) [ 617.321029] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.321029] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 617.321874] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 98f102f8-7a5b-4648-b3ce-a66979a361b5, please check neutron logs for more information. [ 617.321874] env[63372]: Removing descriptor: 19 [ 617.496682] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.109s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.497334] env[63372]: ERROR nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 31028e9f-4749-4a6a-815e-fb2daf8f6c84, please check neutron logs for more information. [ 617.497334] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Traceback (most recent call last): [ 617.497334] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 617.497334] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] self.driver.spawn(context, instance, image_meta, [ 617.497334] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 617.497334] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.497334] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.497334] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] vm_ref = self.build_virtual_machine(instance, [ 617.497334] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.497334] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.497334] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.497683] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] for vif in network_info: [ 617.497683] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 617.497683] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] return self._sync_wrapper(fn, *args, **kwargs) [ 617.497683] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 617.497683] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] self.wait() [ 617.497683] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 617.497683] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] self[:] = self._gt.wait() [ 617.497683] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.497683] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] return self._exit_event.wait() [ 617.497683] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.497683] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] result = hub.switch() [ 617.497683] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.497683] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] return self.greenlet.switch() [ 617.497970] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.497970] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] result = function(*args, **kwargs) [ 617.497970] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.497970] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] return func(*args, **kwargs) [ 617.497970] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.497970] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] raise e [ 617.497970] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.497970] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] nwinfo = self.network_api.allocate_for_instance( [ 617.497970] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.497970] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] created_port_ids = self._update_ports_for_instance( [ 617.497970] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.497970] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] with excutils.save_and_reraise_exception(): [ 617.497970] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.498367] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] self.force_reraise() [ 617.498367] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.498367] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] raise self.value [ 617.498367] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.498367] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] updated_port = self._update_port( [ 617.498367] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.498367] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] _ensure_no_port_binding_failure(port) [ 617.498367] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.498367] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] raise exception.PortBindingFailed(port_id=port['id']) [ 617.498367] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] nova.exception.PortBindingFailed: Binding failed for port 31028e9f-4749-4a6a-815e-fb2daf8f6c84, please check neutron logs for more information. [ 617.498367] env[63372]: ERROR nova.compute.manager [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] [ 617.498645] env[63372]: DEBUG nova.compute.utils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Binding failed for port 31028e9f-4749-4a6a-815e-fb2daf8f6c84, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 617.499337] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.574s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.499521] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.499660] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63372) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 617.499929] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.697s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.503090] env[63372]: DEBUG nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Build of instance fe96690d-4ca0-413b-9e04-09e7f73a0bb7 was re-scheduled: Binding failed for port 31028e9f-4749-4a6a-815e-fb2daf8f6c84, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 617.503594] env[63372]: DEBUG nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 617.503839] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Acquiring lock "refresh_cache-fe96690d-4ca0-413b-9e04-09e7f73a0bb7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.503989] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Acquired lock "refresh_cache-fe96690d-4ca0-413b-9e04-09e7f73a0bb7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.504160] env[63372]: DEBUG nova.network.neutron [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 617.505798] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99494f92-3484-4c54-b1a8-592594caa7ed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.517557] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ae0d08-9de8-4699-9b15-a907ef4e42dc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.535026] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57786115-e029-4f8c-8fad-98435d513681 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.541535] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4e4064-38fd-4018-8360-20e25fe4adc1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.571296] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181583MB free_disk=186GB free_vcpus=48 pci_devices=None {{(pid=63372) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 617.571296] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.582113] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023658, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063127} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.585166] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 617.586979] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7324be5-7727-45eb-a7be-0d3a79886853 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.594458] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023659, 'name': ReconfigVM_Task, 'duration_secs': 0.276267} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.600992] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Reconfigured VM instance instance-00000016 to attach disk [datastore2] 8986e782-4d63-4491-8100-4341a3149812/8986e782-4d63-4491-8100-4341a3149812.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 617.601755] env[63372]: DEBUG nova.network.neutron [-] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.611740] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Reconfiguring VM instance instance-00000017 to attach disk [datastore2] 715408b8-0e76-4b61-a342-b168377cb288/715408b8-0e76-4b61-a342-b168377cb288.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 617.612288] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-23b32afb-10a2-4e10-a2cd-0ac8b5bc0c11 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.614419] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1db5b231-5994-4688-9a1c-12437fd4ae6d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.636501] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for the task: (returnval){ [ 617.636501] env[63372]: value = "task-1023660" [ 617.636501] env[63372]: _type = "Task" [ 617.636501] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.637755] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 617.637755] env[63372]: value = "task-1023661" [ 617.637755] env[63372]: _type = "Task" [ 617.637755] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.649068] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023660, 'name': Rename_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.653321] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023661, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.988699] env[63372]: DEBUG nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 618.011584] env[63372]: DEBUG nova.virt.hardware [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.011824] env[63372]: DEBUG nova.virt.hardware [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.011975] env[63372]: DEBUG nova.virt.hardware [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.012232] env[63372]: DEBUG nova.virt.hardware [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.012382] env[63372]: DEBUG nova.virt.hardware [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.012525] env[63372]: DEBUG nova.virt.hardware [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.012721] env[63372]: DEBUG nova.virt.hardware [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.012938] env[63372]: DEBUG nova.virt.hardware [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.013170] env[63372]: DEBUG nova.virt.hardware [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.013340] env[63372]: DEBUG nova.virt.hardware [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.013508] env[63372]: DEBUG nova.virt.hardware [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.018464] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871e01d2-433b-4505-885c-416af22d28ca {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.026716] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d1ab25-15e1-4720-b3aa-1e5dbbb0ca88 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.040389] env[63372]: ERROR nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 98f102f8-7a5b-4648-b3ce-a66979a361b5, please check neutron logs for more information. [ 618.040389] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Traceback (most recent call last): [ 618.040389] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 618.040389] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] yield resources [ 618.040389] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.040389] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] self.driver.spawn(context, instance, image_meta, [ 618.040389] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 618.040389] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.040389] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.040389] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] vm_ref = self.build_virtual_machine(instance, [ 618.040389] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.040738] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.040738] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.040738] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] for vif in network_info: [ 618.040738] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.040738] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] return self._sync_wrapper(fn, *args, **kwargs) [ 618.040738] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.040738] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] self.wait() [ 618.040738] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.040738] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] self[:] = self._gt.wait() [ 618.040738] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.040738] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] return self._exit_event.wait() [ 618.040738] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 618.040738] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] current.throw(*self._exc) [ 618.041066] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.041066] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] result = function(*args, **kwargs) [ 618.041066] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.041066] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] return func(*args, **kwargs) [ 618.041066] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.041066] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] raise e [ 618.041066] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.041066] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] nwinfo = self.network_api.allocate_for_instance( [ 618.041066] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.041066] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] created_port_ids = self._update_ports_for_instance( [ 618.041066] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.041066] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] with excutils.save_and_reraise_exception(): [ 618.041066] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.041415] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] self.force_reraise() [ 618.041415] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.041415] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] raise self.value [ 618.041415] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.041415] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] updated_port = self._update_port( [ 618.041415] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.041415] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] _ensure_no_port_binding_failure(port) [ 618.041415] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.041415] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] raise exception.PortBindingFailed(port_id=port['id']) [ 618.041415] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] nova.exception.PortBindingFailed: Binding failed for port 98f102f8-7a5b-4648-b3ce-a66979a361b5, please check neutron logs for more information. [ 618.041415] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] [ 618.041415] env[63372]: INFO nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Terminating instance [ 618.042715] env[63372]: DEBUG nova.network.neutron [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.046820] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Acquiring lock "refresh_cache-6aa468f8-cd63-459b-8302-8bc22580b01d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.046978] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Acquired lock "refresh_cache-6aa468f8-cd63-459b-8302-8bc22580b01d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.047147] env[63372]: DEBUG nova.network.neutron [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.100868] env[63372]: DEBUG nova.network.neutron [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.113244] env[63372]: INFO nova.compute.manager [-] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Took 1.04 seconds to deallocate network for instance. [ 618.117567] env[63372]: DEBUG nova.compute.claims [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 618.117761] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.152525] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023660, 'name': Rename_Task, 'duration_secs': 0.131552} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.155560] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 618.156271] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023661, 'name': ReconfigVM_Task, 'duration_secs': 0.274306} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.158665] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0575bbad-1843-4117-8ec0-94f36ac624be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.160817] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Reconfigured VM instance instance-00000017 to attach disk [datastore2] 715408b8-0e76-4b61-a342-b168377cb288/715408b8-0e76-4b61-a342-b168377cb288.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 618.162274] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92bbbff7-e043-41e0-9c3c-b7457c0ece57 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.169142] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for the task: (returnval){ [ 618.169142] env[63372]: value = "task-1023662" [ 618.169142] env[63372]: _type = "Task" [ 618.169142] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.170020] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 618.170020] env[63372]: value = "task-1023663" [ 618.170020] env[63372]: _type = "Task" [ 618.170020] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.181895] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023662, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.185044] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023663, 'name': Rename_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.191380] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "b182294d-2de8-4189-af7f-3e2d2c604a8b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.191602] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "b182294d-2de8-4189-af7f-3e2d2c604a8b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.403105] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "f0c60559-c072-4b61-afe8-03d6c131b307" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.403374] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "f0c60559-c072-4b61-afe8-03d6c131b307" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.413541] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72290fc-0a96-4c87-b096-76003a76f212 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.422129] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d565c9-6d4a-43ac-8dea-ab9d9cab9335 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.453048] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01187c5b-e23d-489c-ad97-820eab466ad7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.460130] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1392e5-0524-40d2-b738-02676a324b16 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.473081] env[63372]: DEBUG nova.compute.provider_tree [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.568410] env[63372]: DEBUG nova.network.neutron [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.606433] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Releasing lock "refresh_cache-fe96690d-4ca0-413b-9e04-09e7f73a0bb7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.606433] env[63372]: DEBUG nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 618.606433] env[63372]: DEBUG nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.606433] env[63372]: DEBUG nova.network.neutron [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 618.626592] env[63372]: DEBUG nova.network.neutron [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.653802] env[63372]: DEBUG nova.network.neutron [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.685068] env[63372]: DEBUG oslo_vmware.api [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023662, 'name': PowerOnVM_Task, 'duration_secs': 0.412701} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.688560] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 618.688763] env[63372]: INFO nova.compute.manager [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Took 9.01 seconds to spawn the instance on the hypervisor. [ 618.689029] env[63372]: DEBUG nova.compute.manager [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 618.689526] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023663, 'name': Rename_Task, 'duration_secs': 0.141908} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.690521] env[63372]: DEBUG nova.compute.manager [req-5f29d429-6c0a-4f5e-afa1-6501c298c453 req-16d8ee05-5fe5-4d8f-a17b-36be4ca89fd3 service nova] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Received event network-changed-98f102f8-7a5b-4648-b3ce-a66979a361b5 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.690695] env[63372]: DEBUG nova.compute.manager [req-5f29d429-6c0a-4f5e-afa1-6501c298c453 req-16d8ee05-5fe5-4d8f-a17b-36be4ca89fd3 service nova] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Refreshing instance network info cache due to event network-changed-98f102f8-7a5b-4648-b3ce-a66979a361b5. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 618.690898] env[63372]: DEBUG oslo_concurrency.lockutils [req-5f29d429-6c0a-4f5e-afa1-6501c298c453 req-16d8ee05-5fe5-4d8f-a17b-36be4ca89fd3 service nova] Acquiring lock "refresh_cache-6aa468f8-cd63-459b-8302-8bc22580b01d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.691564] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c38447-3948-4301-a746-17b592dda32a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.694203] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 618.694618] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e3eb41c-2816-4987-84a6-70c9c12f7ecf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.702828] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 618.702828] env[63372]: value = "task-1023664" [ 618.702828] env[63372]: _type = "Task" [ 618.702828] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.710162] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023664, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.976599] env[63372]: DEBUG nova.scheduler.client.report [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 619.128841] env[63372]: DEBUG nova.network.neutron [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.156087] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Releasing lock "refresh_cache-6aa468f8-cd63-459b-8302-8bc22580b01d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.156579] env[63372]: DEBUG nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 619.156774] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 619.157346] env[63372]: DEBUG oslo_concurrency.lockutils [req-5f29d429-6c0a-4f5e-afa1-6501c298c453 req-16d8ee05-5fe5-4d8f-a17b-36be4ca89fd3 service nova] Acquired lock "refresh_cache-6aa468f8-cd63-459b-8302-8bc22580b01d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.157534] env[63372]: DEBUG nova.network.neutron [req-5f29d429-6c0a-4f5e-afa1-6501c298c453 req-16d8ee05-5fe5-4d8f-a17b-36be4ca89fd3 service nova] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Refreshing network info cache for port 98f102f8-7a5b-4648-b3ce-a66979a361b5 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 619.158531] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f52b6563-d130-4f8d-8b8e-d6eba511f935 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.168827] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f500fff-f384-4d37-aa1d-6417b46f189d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.191758] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6aa468f8-cd63-459b-8302-8bc22580b01d could not be found. [ 619.192789] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 619.192789] env[63372]: INFO nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 619.192789] env[63372]: DEBUG oslo.service.loopingcall [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 619.192941] env[63372]: DEBUG nova.compute.manager [-] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 619.193033] env[63372]: DEBUG nova.network.neutron [-] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 619.215266] env[63372]: DEBUG nova.network.neutron [-] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.215266] env[63372]: INFO nova.compute.manager [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Took 29.85 seconds to build instance. [ 619.221045] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023664, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.482465] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.982s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.483121] env[63372]: ERROR nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5f1c17ad-7301-48b8-8a7d-aacd4717dc74, please check neutron logs for more information. [ 619.483121] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Traceback (most recent call last): [ 619.483121] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.483121] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] self.driver.spawn(context, instance, image_meta, [ 619.483121] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 619.483121] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.483121] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.483121] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] vm_ref = self.build_virtual_machine(instance, [ 619.483121] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.483121] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.483121] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.483435] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] for vif in network_info: [ 619.483435] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.483435] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] return self._sync_wrapper(fn, *args, **kwargs) [ 619.483435] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.483435] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] self.wait() [ 619.483435] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.483435] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] self[:] = self._gt.wait() [ 619.483435] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.483435] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] return self._exit_event.wait() [ 619.483435] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.483435] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] result = hub.switch() [ 619.483435] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.483435] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] return self.greenlet.switch() [ 619.483749] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.483749] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] result = function(*args, **kwargs) [ 619.483749] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.483749] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] return func(*args, **kwargs) [ 619.483749] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.483749] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] raise e [ 619.483749] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.483749] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] nwinfo = self.network_api.allocate_for_instance( [ 619.483749] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 619.483749] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] created_port_ids = self._update_ports_for_instance( [ 619.483749] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 619.483749] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] with excutils.save_and_reraise_exception(): [ 619.483749] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.484041] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] self.force_reraise() [ 619.484041] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.484041] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] raise self.value [ 619.484041] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 619.484041] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] updated_port = self._update_port( [ 619.484041] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.484041] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] _ensure_no_port_binding_failure(port) [ 619.484041] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.484041] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] raise exception.PortBindingFailed(port_id=port['id']) [ 619.484041] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] nova.exception.PortBindingFailed: Binding failed for port 5f1c17ad-7301-48b8-8a7d-aacd4717dc74, please check neutron logs for more information. [ 619.484041] env[63372]: ERROR nova.compute.manager [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] [ 619.484292] env[63372]: DEBUG nova.compute.utils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Binding failed for port 5f1c17ad-7301-48b8-8a7d-aacd4717dc74, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 619.485652] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.747s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.488987] env[63372]: DEBUG nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Build of instance d0310e79-e07a-4581-8cd8-34437cd9ea7d was re-scheduled: Binding failed for port 5f1c17ad-7301-48b8-8a7d-aacd4717dc74, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 619.489444] env[63372]: DEBUG nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 619.489672] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "refresh_cache-d0310e79-e07a-4581-8cd8-34437cd9ea7d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.489810] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquired lock "refresh_cache-d0310e79-e07a-4581-8cd8-34437cd9ea7d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.489960] env[63372]: DEBUG nova.network.neutron [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 619.556929] env[63372]: DEBUG nova.compute.manager [None req-8bab402b-b625-48d1-ba10-43e8092a7205 tempest-ServerDiagnosticsV248Test-1513623816 tempest-ServerDiagnosticsV248Test-1513623816-project-admin] [instance: 8986e782-4d63-4491-8100-4341a3149812] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 619.558072] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06608436-4b86-4a90-8785-a025f6724704 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.566096] env[63372]: INFO nova.compute.manager [None req-8bab402b-b625-48d1-ba10-43e8092a7205 tempest-ServerDiagnosticsV248Test-1513623816 tempest-ServerDiagnosticsV248Test-1513623816-project-admin] [instance: 8986e782-4d63-4491-8100-4341a3149812] Retrieving diagnostics [ 619.566900] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a386ba-fd6e-4c79-979f-2b6e487258b6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.631859] env[63372]: INFO nova.compute.manager [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] [instance: fe96690d-4ca0-413b-9e04-09e7f73a0bb7] Took 1.02 seconds to deallocate network for instance. [ 619.679825] env[63372]: DEBUG nova.network.neutron [req-5f29d429-6c0a-4f5e-afa1-6501c298c453 req-16d8ee05-5fe5-4d8f-a17b-36be4ca89fd3 service nova] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.716719] env[63372]: DEBUG nova.network.neutron [-] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.718461] env[63372]: DEBUG oslo_vmware.api [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023664, 'name': PowerOnVM_Task, 'duration_secs': 1.001633} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.718889] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a598c57b-850a-409c-a357-563cc2d20e32 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Lock "8986e782-4d63-4491-8100-4341a3149812" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.867s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.719447] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 619.720027] env[63372]: INFO nova.compute.manager [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Took 7.48 seconds to spawn the instance on the hypervisor. [ 619.720027] env[63372]: DEBUG nova.compute.manager [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 619.721804] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffaf1a13-a804-42b7-9559-ae8e96ffe50f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.768150] env[63372]: DEBUG nova.network.neutron [req-5f29d429-6c0a-4f5e-afa1-6501c298c453 req-16d8ee05-5fe5-4d8f-a17b-36be4ca89fd3 service nova] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.018939] env[63372]: DEBUG nova.network.neutron [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.081391] env[63372]: DEBUG nova.network.neutron [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.219619] env[63372]: INFO nova.compute.manager [-] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Took 1.03 seconds to deallocate network for instance. [ 620.222448] env[63372]: DEBUG nova.compute.claims [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 620.222628] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.222920] env[63372]: DEBUG nova.compute.manager [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 620.240914] env[63372]: INFO nova.compute.manager [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Took 28.75 seconds to build instance. [ 620.272155] env[63372]: DEBUG oslo_concurrency.lockutils [req-5f29d429-6c0a-4f5e-afa1-6501c298c453 req-16d8ee05-5fe5-4d8f-a17b-36be4ca89fd3 service nova] Releasing lock "refresh_cache-6aa468f8-cd63-459b-8302-8bc22580b01d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.272418] env[63372]: DEBUG nova.compute.manager [req-5f29d429-6c0a-4f5e-afa1-6501c298c453 req-16d8ee05-5fe5-4d8f-a17b-36be4ca89fd3 service nova] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Received event network-vif-deleted-98f102f8-7a5b-4648-b3ce-a66979a361b5 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 620.377702] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077d248b-34ae-4aae-a1aa-bc316a192674 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.391060] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21de7fd5-0a77-4fa2-a2d0-bb0909b54e79 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.422234] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3dc49b7-d4de-48f1-acd8-0b582c48fa35 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.429543] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080f0155-f495-4358-91ab-7afe3f5b106b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.442687] env[63372]: DEBUG nova.compute.provider_tree [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.585809] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Releasing lock "refresh_cache-d0310e79-e07a-4581-8cd8-34437cd9ea7d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.586047] env[63372]: DEBUG nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 620.586216] env[63372]: DEBUG nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.586380] env[63372]: DEBUG nova.network.neutron [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 620.601324] env[63372]: DEBUG nova.network.neutron [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.657542] env[63372]: INFO nova.scheduler.client.report [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Deleted allocations for instance fe96690d-4ca0-413b-9e04-09e7f73a0bb7 [ 620.741619] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.743202] env[63372]: DEBUG oslo_concurrency.lockutils [None req-366b8038-505b-4b2f-abd2-96b5924862b5 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Lock "715408b8-0e76-4b61-a342-b168377cb288" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.052s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.885947] env[63372]: INFO nova.compute.manager [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Rebuilding instance [ 620.926731] env[63372]: DEBUG nova.compute.manager [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 620.927889] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f874c5-e045-40d4-842c-befe5c30906e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.945892] env[63372]: DEBUG nova.scheduler.client.report [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.103944] env[63372]: DEBUG nova.network.neutron [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.165947] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75c7a5d2-70eb-484c-8c25-a851b5dfb00e tempest-ServerDiagnosticsTest-99653979 tempest-ServerDiagnosticsTest-99653979-project-member] Lock "fe96690d-4ca0-413b-9e04-09e7f73a0bb7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.090s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.246555] env[63372]: DEBUG nova.compute.manager [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 621.439009] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 621.439673] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7af05d91-28ea-4e73-9b52-490ed68b204c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.446818] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 621.446818] env[63372]: value = "task-1023665" [ 621.446818] env[63372]: _type = "Task" [ 621.446818] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.451031] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.965s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.451359] env[63372]: ERROR nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 852395e7-45d6-4d16-8b05-2144329e6dc4, please check neutron logs for more information. [ 621.451359] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Traceback (most recent call last): [ 621.451359] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.451359] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] self.driver.spawn(context, instance, image_meta, [ 621.451359] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 621.451359] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.451359] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.451359] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] vm_ref = self.build_virtual_machine(instance, [ 621.451359] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.451359] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.451359] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.451641] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] for vif in network_info: [ 621.451641] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.451641] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] return self._sync_wrapper(fn, *args, **kwargs) [ 621.451641] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.451641] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] self.wait() [ 621.451641] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.451641] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] self[:] = self._gt.wait() [ 621.451641] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.451641] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] return self._exit_event.wait() [ 621.451641] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.451641] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] result = hub.switch() [ 621.451641] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.451641] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] return self.greenlet.switch() [ 621.451925] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.451925] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] result = function(*args, **kwargs) [ 621.451925] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.451925] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] return func(*args, **kwargs) [ 621.451925] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.451925] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] raise e [ 621.451925] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.451925] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] nwinfo = self.network_api.allocate_for_instance( [ 621.451925] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 621.451925] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] created_port_ids = self._update_ports_for_instance( [ 621.451925] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 621.451925] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] with excutils.save_and_reraise_exception(): [ 621.451925] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.452218] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] self.force_reraise() [ 621.452218] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.452218] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] raise self.value [ 621.452218] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 621.452218] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] updated_port = self._update_port( [ 621.452218] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.452218] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] _ensure_no_port_binding_failure(port) [ 621.452218] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.452218] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] raise exception.PortBindingFailed(port_id=port['id']) [ 621.452218] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] nova.exception.PortBindingFailed: Binding failed for port 852395e7-45d6-4d16-8b05-2144329e6dc4, please check neutron logs for more information. [ 621.452218] env[63372]: ERROR nova.compute.manager [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] [ 621.452462] env[63372]: DEBUG nova.compute.utils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Binding failed for port 852395e7-45d6-4d16-8b05-2144329e6dc4, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 621.453228] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.880s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.456939] env[63372]: DEBUG nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Build of instance 6a6268e4-0539-45cf-9335-cde31473d1f9 was re-scheduled: Binding failed for port 852395e7-45d6-4d16-8b05-2144329e6dc4, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 621.457401] env[63372]: DEBUG nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 621.457650] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Acquiring lock "refresh_cache-6a6268e4-0539-45cf-9335-cde31473d1f9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.457820] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Acquired lock "refresh_cache-6a6268e4-0539-45cf-9335-cde31473d1f9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.457980] env[63372]: DEBUG nova.network.neutron [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 621.461935] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023665, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.608671] env[63372]: INFO nova.compute.manager [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: d0310e79-e07a-4581-8cd8-34437cd9ea7d] Took 1.02 seconds to deallocate network for instance. [ 621.669271] env[63372]: DEBUG nova.compute.manager [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 621.767298] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.957691] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023665, 'name': PowerOffVM_Task, 'duration_secs': 0.111239} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.957781] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 621.957946] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 621.958716] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f429ad8e-fd69-4894-8694-adaf65d7a523 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.970139] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 621.970664] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97ba6824-9dab-4e27-808d-76f995b9bf59 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.985994] env[63372]: DEBUG nova.network.neutron [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.996065] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 621.996467] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 621.996784] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Deleting the datastore file [datastore2] 715408b8-0e76-4b61-a342-b168377cb288 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 621.997184] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4b3ec94-a0a0-4e08-8088-448a75ef61c3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.004166] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 622.004166] env[63372]: value = "task-1023667" [ 622.004166] env[63372]: _type = "Task" [ 622.004166] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.015849] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.111381] env[63372]: DEBUG nova.network.neutron [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.191965] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.410948] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5726ae38-d285-476a-980a-e39d31ea4f6c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.418662] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54385cbb-f1fc-4d7d-84d0-fba7f6ccf1d7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.448713] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d14d10-bdd1-4c57-903b-91e2455a5155 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.457289] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee228097-ef76-484c-ae37-86428bb540a9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.471702] env[63372]: DEBUG nova.compute.provider_tree [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.514594] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023667, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096528} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.514890] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 622.515084] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 622.515315] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 622.616646] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Releasing lock "refresh_cache-6a6268e4-0539-45cf-9335-cde31473d1f9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.616889] env[63372]: DEBUG nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 622.617092] env[63372]: DEBUG nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.617264] env[63372]: DEBUG nova.network.neutron [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 622.637377] env[63372]: DEBUG nova.network.neutron [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.638398] env[63372]: INFO nova.scheduler.client.report [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Deleted allocations for instance d0310e79-e07a-4581-8cd8-34437cd9ea7d [ 622.975635] env[63372]: DEBUG nova.scheduler.client.report [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.145346] env[63372]: DEBUG nova.network.neutron [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.151037] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1146ad1c-5f20-4358-a170-ad07ca443c4e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "d0310e79-e07a-4581-8cd8-34437cd9ea7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.379s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.480435] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.027s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.480966] env[63372]: ERROR nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5eaf8da0-4d11-4f42-abe3-d5f48f7784cd, please check neutron logs for more information. [ 623.480966] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Traceback (most recent call last): [ 623.480966] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.480966] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] self.driver.spawn(context, instance, image_meta, [ 623.480966] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 623.480966] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.480966] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.480966] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] vm_ref = self.build_virtual_machine(instance, [ 623.480966] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.480966] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.480966] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.481299] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] for vif in network_info: [ 623.481299] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.481299] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] return self._sync_wrapper(fn, *args, **kwargs) [ 623.481299] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.481299] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] self.wait() [ 623.481299] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.481299] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] self[:] = self._gt.wait() [ 623.481299] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.481299] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] return self._exit_event.wait() [ 623.481299] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.481299] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] result = hub.switch() [ 623.481299] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.481299] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] return self.greenlet.switch() [ 623.481596] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.481596] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] result = function(*args, **kwargs) [ 623.481596] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.481596] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] return func(*args, **kwargs) [ 623.481596] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.481596] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] raise e [ 623.481596] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.481596] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] nwinfo = self.network_api.allocate_for_instance( [ 623.481596] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.481596] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] created_port_ids = self._update_ports_for_instance( [ 623.481596] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.481596] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] with excutils.save_and_reraise_exception(): [ 623.481596] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.481888] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] self.force_reraise() [ 623.481888] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.481888] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] raise self.value [ 623.481888] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.481888] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] updated_port = self._update_port( [ 623.481888] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.481888] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] _ensure_no_port_binding_failure(port) [ 623.481888] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.481888] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] raise exception.PortBindingFailed(port_id=port['id']) [ 623.481888] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] nova.exception.PortBindingFailed: Binding failed for port 5eaf8da0-4d11-4f42-abe3-d5f48f7784cd, please check neutron logs for more information. [ 623.481888] env[63372]: ERROR nova.compute.manager [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] [ 623.482155] env[63372]: DEBUG nova.compute.utils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Binding failed for port 5eaf8da0-4d11-4f42-abe3-d5f48f7784cd, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 623.483157] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.042s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.484497] env[63372]: INFO nova.compute.claims [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 623.487489] env[63372]: DEBUG nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Build of instance 96c3ce22-9424-4f7f-8810-ee7cfaed0193 was re-scheduled: Binding failed for port 5eaf8da0-4d11-4f42-abe3-d5f48f7784cd, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 623.487849] env[63372]: DEBUG nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 623.488078] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Acquiring lock "refresh_cache-96c3ce22-9424-4f7f-8810-ee7cfaed0193" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.488224] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Acquired lock "refresh_cache-96c3ce22-9424-4f7f-8810-ee7cfaed0193" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.488424] env[63372]: DEBUG nova.network.neutron [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 623.551018] env[63372]: DEBUG nova.virt.hardware [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 623.551018] env[63372]: DEBUG nova.virt.hardware [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 623.551018] env[63372]: DEBUG nova.virt.hardware [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.551018] env[63372]: DEBUG nova.virt.hardware [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 623.551319] env[63372]: DEBUG nova.virt.hardware [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.551319] env[63372]: DEBUG nova.virt.hardware [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 623.551319] env[63372]: DEBUG nova.virt.hardware [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 623.551712] env[63372]: DEBUG nova.virt.hardware [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 623.552031] env[63372]: DEBUG nova.virt.hardware [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 623.552324] env[63372]: DEBUG nova.virt.hardware [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 623.552621] env[63372]: DEBUG nova.virt.hardware [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.555032] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a0891a-0e26-44d0-92b6-5dbeb6578260 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.562357] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7ddef1-7dac-4cc6-a4fb-0b77dc3e2149 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.577954] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Instance VIF info [] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 623.584392] env[63372]: DEBUG oslo.service.loopingcall [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 623.584545] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 623.585034] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-337b7505-8352-4bc6-a69f-cbd2496496d4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.601589] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 623.601589] env[63372]: value = "task-1023668" [ 623.601589] env[63372]: _type = "Task" [ 623.601589] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.609737] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023668, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.650778] env[63372]: INFO nova.compute.manager [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] [instance: 6a6268e4-0539-45cf-9335-cde31473d1f9] Took 1.03 seconds to deallocate network for instance. [ 623.656029] env[63372]: DEBUG nova.compute.manager [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 624.014281] env[63372]: DEBUG nova.network.neutron [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.112133] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023668, 'name': CreateVM_Task, 'duration_secs': 0.27858} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.113371] env[63372]: DEBUG nova.network.neutron [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.114868] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 624.116229] env[63372]: DEBUG oslo_vmware.service [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede506cd-141d-44d3-871f-16046c0f71b3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.124786] env[63372]: DEBUG oslo_concurrency.lockutils [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.125514] env[63372]: DEBUG oslo_concurrency.lockutils [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.129502] env[63372]: DEBUG oslo_concurrency.lockutils [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 624.129502] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab28877f-f126-4719-95af-a53bb914dd9b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.136034] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 624.136034] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5296a2dd-20c6-cd61-0c21-dc1bf4eb894a" [ 624.136034] env[63372]: _type = "Task" [ 624.136034] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.149154] env[63372]: DEBUG oslo_concurrency.lockutils [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.149937] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 624.149937] env[63372]: DEBUG oslo_concurrency.lockutils [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.149937] env[63372]: DEBUG oslo_concurrency.lockutils [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.150334] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 624.151058] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-174a61db-80ad-4473-81c7-e36c45012ee6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.169761] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 624.170026] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 624.171050] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d937af-8a7c-4719-ac28-f4b47cfe1c60 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.183294] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b67313c-68f7-45ad-ab21-a485050085ad {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.189253] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 624.189253] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52844b92-2e42-ad0f-443f-97cdb7edee2b" [ 624.189253] env[63372]: _type = "Task" [ 624.189253] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.195380] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.198880] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52844b92-2e42-ad0f-443f-97cdb7edee2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.618024] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Releasing lock "refresh_cache-96c3ce22-9424-4f7f-8810-ee7cfaed0193" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.618024] env[63372]: DEBUG nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 624.618024] env[63372]: DEBUG nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.618024] env[63372]: DEBUG nova.network.neutron [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 624.641513] env[63372]: DEBUG nova.network.neutron [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.688961] env[63372]: INFO nova.scheduler.client.report [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Deleted allocations for instance 6a6268e4-0539-45cf-9335-cde31473d1f9 [ 624.712404] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Preparing fetch location {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 624.712856] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Creating directory with path [datastore1] vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d/d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 624.713248] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a521f3d1-7d60-453d-8de5-f9694d452555 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.755718] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Created directory with path [datastore1] vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d/d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 624.755718] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Fetch image to [datastore1] vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 624.755718] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Downloading image file data d7ae1717-77b2-47f6-9acd-b27bac221f7c to [datastore1] vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk on the data store datastore1 {{(pid=63372) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 624.756426] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd83d7f-070c-4aec-958e-3414d613b246 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.770018] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec8208b-04db-439a-8142-f34b2aa71071 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.784058] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d1c3b1-bdde-438b-81eb-10966aa38ed2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.828815] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3e4e96-5807-4b17-9c5b-0ca473151fd8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.835896] env[63372]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-29a676b2-b827-4985-bae2-9a97919848ec {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.863048] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Downloading image file data d7ae1717-77b2-47f6-9acd-b27bac221f7c to the data store datastore1 {{(pid=63372) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 624.869924] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "8283b736-ad02-4082-97b7-561bd5c5da93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.871109] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "8283b736-ad02-4082-97b7-561bd5c5da93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.931278] env[63372]: DEBUG oslo_vmware.rw_handles [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63372) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 625.040016] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2f30f4-28c6-4e3a-8522-232fe4b8a465 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.053388] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7eed9bb-f62d-4c79-86ba-4f0b9fb75e53 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.087522] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d158ad0-5a05-4f84-aafc-f8bc8768e57d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.096035] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5229b48-91fa-4a13-9f0e-90b3f0302b27 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.112824] env[63372]: DEBUG nova.compute.provider_tree [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.144331] env[63372]: DEBUG nova.network.neutron [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.205865] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f4cc62ba-5b07-46ec-aa25-cc6fa9453382 tempest-ServersWithSpecificFlavorTestJSON-1492193662 tempest-ServersWithSpecificFlavorTestJSON-1492193662-project-member] Lock "6a6268e4-0539-45cf-9335-cde31473d1f9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.090s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.618887] env[63372]: DEBUG nova.scheduler.client.report [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.648577] env[63372]: INFO nova.compute.manager [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] [instance: 96c3ce22-9424-4f7f-8810-ee7cfaed0193] Took 1.03 seconds to deallocate network for instance. [ 625.709313] env[63372]: DEBUG nova.compute.manager [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.760977] env[63372]: DEBUG oslo_vmware.rw_handles [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Completed reading data from the image iterator. {{(pid=63372) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 625.761366] env[63372]: DEBUG oslo_vmware.rw_handles [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 625.837739] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Downloaded image file data d7ae1717-77b2-47f6-9acd-b27bac221f7c to vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk on the data store datastore1 {{(pid=63372) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 625.839842] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Caching image {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 625.840040] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Copying Virtual Disk [datastore1] vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk to [datastore1] vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 625.840353] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a54a4fd-d657-4d55-ac6a-64d703fc9d06 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.848272] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 625.848272] env[63372]: value = "task-1023669" [ 625.848272] env[63372]: _type = "Task" [ 625.848272] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.857060] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023669, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.125346] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.642s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.128016] env[63372]: DEBUG nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 626.132746] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.527s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.236571] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.358684] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023669, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.639014] env[63372]: DEBUG nova.compute.utils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 626.640703] env[63372]: DEBUG nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 626.640868] env[63372]: DEBUG nova.network.neutron [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 626.690735] env[63372]: INFO nova.scheduler.client.report [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Deleted allocations for instance 96c3ce22-9424-4f7f-8810-ee7cfaed0193 [ 626.705666] env[63372]: DEBUG nova.policy [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5900eb23816b4a809a4ea18c85a3de34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ac244ce6328437f9650b4de78917b8f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 626.860499] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023669, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.71602} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.860961] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Copied Virtual Disk [datastore1] vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk to [datastore1] vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 626.862012] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Deleting the datastore file [datastore1] vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d/d7ae1717-77b2-47f6-9acd-b27bac221f7c/tmp-sparse.vmdk {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 626.862012] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c05bbd19-55d3-4c5b-b2d0-250e84a99ac2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.869211] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 626.869211] env[63372]: value = "task-1023670" [ 626.869211] env[63372]: _type = "Task" [ 626.869211] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.884833] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023670, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.070161] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d4ada1-1394-40ac-be65-0e264e643709 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.074450] env[63372]: DEBUG nova.network.neutron [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Successfully created port: 61938f95-e588-4bdf-b61a-0795bd8279e3 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 627.082159] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab982d4-5595-40d8-a591-79b82c9c9328 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.113370] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b1b4436-891e-4c25-bd6d-e2e979283372 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.121832] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ba1b39-9305-434f-b50b-a4235c4ce650 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.135750] env[63372]: DEBUG nova.compute.provider_tree [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.148220] env[63372]: DEBUG nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 627.198660] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cc30d8d-695d-4f11-8d6d-1786bad5f4ae tempest-ServerExternalEventsTest-97610754 tempest-ServerExternalEventsTest-97610754-project-member] Lock "96c3ce22-9424-4f7f-8810-ee7cfaed0193" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.218s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.381187] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023670, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025878} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.381309] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 627.381442] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Moving file from [datastore1] vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d/d7ae1717-77b2-47f6-9acd-b27bac221f7c to [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c. {{(pid=63372) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 627.381689] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-12bf4723-d65a-4723-8ef8-e0bae8c5e442 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.388427] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 627.388427] env[63372]: value = "task-1023671" [ 627.388427] env[63372]: _type = "Task" [ 627.388427] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.396329] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023671, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.558893] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "93a5d948-0629-4f53-a681-858d519acfa7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.559345] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "93a5d948-0629-4f53-a681-858d519acfa7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.641793] env[63372]: DEBUG nova.scheduler.client.report [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.701371] env[63372]: DEBUG nova.compute.manager [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 627.900861] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023671, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024872} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.901831] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] File moved {{(pid=63372) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 627.902091] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Cleaning up location [datastore1] vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 627.902266] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Deleting the datastore file [datastore1] vmware_temp/8b62a9f9-b516-4976-b3f7-bbe1ab4a7e4d {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 627.902533] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fa35eb43-b4bb-4d5c-a427-bfc20cd83598 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.911429] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 627.911429] env[63372]: value = "task-1023672" [ 627.911429] env[63372]: _type = "Task" [ 627.911429] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.919264] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023672, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.149174] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.018s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.150466] env[63372]: ERROR nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 40531f78-08e2-4c40-bfa7-c896e07e4538, please check neutron logs for more information. [ 628.150466] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Traceback (most recent call last): [ 628.150466] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.150466] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] self.driver.spawn(context, instance, image_meta, [ 628.150466] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 628.150466] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.150466] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.150466] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] vm_ref = self.build_virtual_machine(instance, [ 628.150466] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.150466] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.150466] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.150746] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] for vif in network_info: [ 628.150746] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.150746] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] return self._sync_wrapper(fn, *args, **kwargs) [ 628.150746] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.150746] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] self.wait() [ 628.150746] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.150746] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] self[:] = self._gt.wait() [ 628.150746] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.150746] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] return self._exit_event.wait() [ 628.150746] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.150746] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] result = hub.switch() [ 628.150746] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.150746] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] return self.greenlet.switch() [ 628.151041] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.151041] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] result = function(*args, **kwargs) [ 628.151041] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.151041] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] return func(*args, **kwargs) [ 628.151041] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.151041] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] raise e [ 628.151041] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.151041] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] nwinfo = self.network_api.allocate_for_instance( [ 628.151041] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.151041] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] created_port_ids = self._update_ports_for_instance( [ 628.151041] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.151041] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] with excutils.save_and_reraise_exception(): [ 628.151041] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.151366] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] self.force_reraise() [ 628.151366] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.151366] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] raise self.value [ 628.151366] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.151366] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] updated_port = self._update_port( [ 628.151366] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.151366] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] _ensure_no_port_binding_failure(port) [ 628.151366] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.151366] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] raise exception.PortBindingFailed(port_id=port['id']) [ 628.151366] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] nova.exception.PortBindingFailed: Binding failed for port 40531f78-08e2-4c40-bfa7-c896e07e4538, please check neutron logs for more information. [ 628.151366] env[63372]: ERROR nova.compute.manager [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] [ 628.152217] env[63372]: DEBUG nova.compute.utils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Binding failed for port 40531f78-08e2-4c40-bfa7-c896e07e4538, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 628.153368] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.582s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.155796] env[63372]: DEBUG nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Build of instance c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08 was re-scheduled: Binding failed for port 40531f78-08e2-4c40-bfa7-c896e07e4538, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 628.156434] env[63372]: DEBUG nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 628.156714] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Acquiring lock "refresh_cache-c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.156939] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Acquired lock "refresh_cache-c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.157191] env[63372]: DEBUG nova.network.neutron [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 628.161404] env[63372]: DEBUG nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 628.168801] env[63372]: DEBUG nova.compute.manager [req-9f42684c-89dd-4a10-bd30-4af4bcd2f366 req-75e2baa2-c830-4a22-b9fb-43b7c0e82bb9 service nova] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Received event network-changed-61938f95-e588-4bdf-b61a-0795bd8279e3 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 628.169045] env[63372]: DEBUG nova.compute.manager [req-9f42684c-89dd-4a10-bd30-4af4bcd2f366 req-75e2baa2-c830-4a22-b9fb-43b7c0e82bb9 service nova] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Refreshing instance network info cache due to event network-changed-61938f95-e588-4bdf-b61a-0795bd8279e3. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 628.169310] env[63372]: DEBUG oslo_concurrency.lockutils [req-9f42684c-89dd-4a10-bd30-4af4bcd2f366 req-75e2baa2-c830-4a22-b9fb-43b7c0e82bb9 service nova] Acquiring lock "refresh_cache-ae86a977-0a9a-46e5-9072-cfb772324393" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.169849] env[63372]: DEBUG oslo_concurrency.lockutils [req-9f42684c-89dd-4a10-bd30-4af4bcd2f366 req-75e2baa2-c830-4a22-b9fb-43b7c0e82bb9 service nova] Acquired lock "refresh_cache-ae86a977-0a9a-46e5-9072-cfb772324393" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.170136] env[63372]: DEBUG nova.network.neutron [req-9f42684c-89dd-4a10-bd30-4af4bcd2f366 req-75e2baa2-c830-4a22-b9fb-43b7c0e82bb9 service nova] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Refreshing network info cache for port 61938f95-e588-4bdf-b61a-0795bd8279e3 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 628.201756] env[63372]: DEBUG nova.virt.hardware [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 628.202044] env[63372]: DEBUG nova.virt.hardware [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 628.202164] env[63372]: DEBUG nova.virt.hardware [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 628.202350] env[63372]: DEBUG nova.virt.hardware [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 628.202575] env[63372]: DEBUG nova.virt.hardware [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 628.202780] env[63372]: DEBUG nova.virt.hardware [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 628.202974] env[63372]: DEBUG nova.virt.hardware [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 628.203393] env[63372]: DEBUG nova.virt.hardware [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 628.204450] env[63372]: DEBUG nova.virt.hardware [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 628.204450] env[63372]: DEBUG nova.virt.hardware [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 628.204450] env[63372]: DEBUG nova.virt.hardware [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.205695] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6020dd2f-4e9a-46c3-8620-149c37ba8cbf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.222051] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd4a7b0-e267-4951-892c-03260fa98b2a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.241278] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.353751] env[63372]: ERROR nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 61938f95-e588-4bdf-b61a-0795bd8279e3, please check neutron logs for more information. [ 628.353751] env[63372]: ERROR nova.compute.manager Traceback (most recent call last): [ 628.353751] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.353751] env[63372]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 628.353751] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.353751] env[63372]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 628.353751] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.353751] env[63372]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 628.353751] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.353751] env[63372]: ERROR nova.compute.manager self.force_reraise() [ 628.353751] env[63372]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.353751] env[63372]: ERROR nova.compute.manager raise self.value [ 628.353751] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.353751] env[63372]: ERROR nova.compute.manager updated_port = self._update_port( [ 628.353751] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.353751] env[63372]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 628.354289] env[63372]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.354289] env[63372]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 628.354289] env[63372]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 61938f95-e588-4bdf-b61a-0795bd8279e3, please check neutron logs for more information. [ 628.354289] env[63372]: ERROR nova.compute.manager [ 628.354289] env[63372]: Traceback (most recent call last): [ 628.354289] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 628.354289] env[63372]: listener.cb(fileno) [ 628.354289] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.354289] env[63372]: result = function(*args, **kwargs) [ 628.354289] env[63372]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.354289] env[63372]: return func(*args, **kwargs) [ 628.354289] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.354289] env[63372]: raise e [ 628.354289] env[63372]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.354289] env[63372]: nwinfo = self.network_api.allocate_for_instance( [ 628.354289] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.354289] env[63372]: created_port_ids = self._update_ports_for_instance( [ 628.354289] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.354289] env[63372]: with excutils.save_and_reraise_exception(): [ 628.354289] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.354289] env[63372]: self.force_reraise() [ 628.354289] env[63372]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.354289] env[63372]: raise self.value [ 628.354289] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.354289] env[63372]: updated_port = self._update_port( [ 628.354289] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.354289] env[63372]: _ensure_no_port_binding_failure(port) [ 628.354289] env[63372]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.354289] env[63372]: raise exception.PortBindingFailed(port_id=port['id']) [ 628.354961] env[63372]: nova.exception.PortBindingFailed: Binding failed for port 61938f95-e588-4bdf-b61a-0795bd8279e3, please check neutron logs for more information. [ 628.354961] env[63372]: Removing descriptor: 19 [ 628.355021] env[63372]: ERROR nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 61938f95-e588-4bdf-b61a-0795bd8279e3, please check neutron logs for more information. [ 628.355021] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Traceback (most recent call last): [ 628.355021] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 628.355021] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] yield resources [ 628.355021] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.355021] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] self.driver.spawn(context, instance, image_meta, [ 628.355021] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 628.355021] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.355021] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.355021] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] vm_ref = self.build_virtual_machine(instance, [ 628.355021] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.355321] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.355321] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.355321] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] for vif in network_info: [ 628.355321] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.355321] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] return self._sync_wrapper(fn, *args, **kwargs) [ 628.355321] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.355321] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] self.wait() [ 628.355321] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.355321] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] self[:] = self._gt.wait() [ 628.355321] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.355321] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] return self._exit_event.wait() [ 628.355321] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.355321] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] result = hub.switch() [ 628.355657] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.355657] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] return self.greenlet.switch() [ 628.355657] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.355657] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] result = function(*args, **kwargs) [ 628.355657] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.355657] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] return func(*args, **kwargs) [ 628.355657] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.355657] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] raise e [ 628.355657] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.355657] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] nwinfo = self.network_api.allocate_for_instance( [ 628.355657] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.355657] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] created_port_ids = self._update_ports_for_instance( [ 628.355657] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.356078] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] with excutils.save_and_reraise_exception(): [ 628.356078] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.356078] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] self.force_reraise() [ 628.356078] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.356078] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] raise self.value [ 628.356078] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.356078] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] updated_port = self._update_port( [ 628.356078] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.356078] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] _ensure_no_port_binding_failure(port) [ 628.356078] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.356078] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] raise exception.PortBindingFailed(port_id=port['id']) [ 628.356078] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] nova.exception.PortBindingFailed: Binding failed for port 61938f95-e588-4bdf-b61a-0795bd8279e3, please check neutron logs for more information. [ 628.356078] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] [ 628.356427] env[63372]: INFO nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Terminating instance [ 628.358130] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Acquiring lock "refresh_cache-ae86a977-0a9a-46e5-9072-cfb772324393" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.425033] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023672, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.029192} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.425815] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 628.426984] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4834e644-3ac7-4c36-af02-0d3d9b381bfc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.435506] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 628.435506] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5260ed81-0f01-fd9a-2c25-238e3844a356" [ 628.435506] env[63372]: _type = "Task" [ 628.435506] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.445479] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5260ed81-0f01-fd9a-2c25-238e3844a356, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.683659] env[63372]: DEBUG nova.network.neutron [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.702394] env[63372]: DEBUG nova.network.neutron [req-9f42684c-89dd-4a10-bd30-4af4bcd2f366 req-75e2baa2-c830-4a22-b9fb-43b7c0e82bb9 service nova] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.950772] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5260ed81-0f01-fd9a-2c25-238e3844a356, 'name': SearchDatastore_Task, 'duration_secs': 0.009544} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.951054] env[63372]: DEBUG oslo_concurrency.lockutils [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.951311] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 715408b8-0e76-4b61-a342-b168377cb288/715408b8-0e76-4b61-a342-b168377cb288.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 628.951573] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a4a1970e-75f0-4ef9-9f42-5b4f6565b108 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.961432] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 628.961432] env[63372]: value = "task-1023673" [ 628.961432] env[63372]: _type = "Task" [ 628.961432] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.974154] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023673, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.015507] env[63372]: DEBUG nova.network.neutron [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.020947] env[63372]: DEBUG nova.network.neutron [req-9f42684c-89dd-4a10-bd30-4af4bcd2f366 req-75e2baa2-c830-4a22-b9fb-43b7c0e82bb9 service nova] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.475152] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023673, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495654} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.475656] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 715408b8-0e76-4b61-a342-b168377cb288/715408b8-0e76-4b61-a342-b168377cb288.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 629.475778] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 629.476083] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30e33eb4-c44f-4393-8a65-8d73e707ecd3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.482887] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 629.482887] env[63372]: value = "task-1023674" [ 629.482887] env[63372]: _type = "Task" [ 629.482887] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.496769] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023674, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.517726] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Releasing lock "refresh_cache-c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.518028] env[63372]: DEBUG nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 629.518227] env[63372]: DEBUG nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.518459] env[63372]: DEBUG nova.network.neutron [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 629.525409] env[63372]: DEBUG oslo_concurrency.lockutils [req-9f42684c-89dd-4a10-bd30-4af4bcd2f366 req-75e2baa2-c830-4a22-b9fb-43b7c0e82bb9 service nova] Releasing lock "refresh_cache-ae86a977-0a9a-46e5-9072-cfb772324393" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.525780] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Acquired lock "refresh_cache-ae86a977-0a9a-46e5-9072-cfb772324393" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.526055] env[63372]: DEBUG nova.network.neutron [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 629.546152] env[63372]: DEBUG nova.network.neutron [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.697147] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 629.697464] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 8986e782-4d63-4491-8100-4341a3149812 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 629.697464] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 715408b8-0e76-4b61-a342-b168377cb288 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 629.697584] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance bd8cf6ba-8039-4c02-bda2-13c7d14f8903 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 629.697750] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 6aa468f8-cd63-459b-8302-8bc22580b01d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 629.697865] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance ae86a977-0a9a-46e5-9072-cfb772324393 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 629.992436] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023674, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066319} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.992866] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 629.994136] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70e9ba3-951c-4d5c-9d01-cb83fd1e0762 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.023890] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Reconfiguring VM instance instance-00000017 to attach disk [datastore1] 715408b8-0e76-4b61-a342-b168377cb288/715408b8-0e76-4b61-a342-b168377cb288.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 630.024262] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-984487e3-ea44-42fe-8fe7-9fd78ee59e43 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.051021] env[63372]: DEBUG nova.network.neutron [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.051021] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 630.051021] env[63372]: value = "task-1023675" [ 630.051021] env[63372]: _type = "Task" [ 630.051021] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.059852] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023675, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.060761] env[63372]: DEBUG nova.network.neutron [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.163922] env[63372]: DEBUG nova.network.neutron [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.190698] env[63372]: DEBUG nova.compute.manager [None req-1d202612-1852-42ac-ae9b-9955f4949601 tempest-ServerDiagnosticsV248Test-1513623816 tempest-ServerDiagnosticsV248Test-1513623816-project-admin] [instance: 8986e782-4d63-4491-8100-4341a3149812] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 630.192147] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deeca8c7-8346-4019-ae63-3d597dfbb2da {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.199586] env[63372]: INFO nova.compute.manager [None req-1d202612-1852-42ac-ae9b-9955f4949601 tempest-ServerDiagnosticsV248Test-1513623816 tempest-ServerDiagnosticsV248Test-1513623816-project-admin] [instance: 8986e782-4d63-4491-8100-4341a3149812] Retrieving diagnostics [ 630.200616] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 6d4502b7-b9df-4fef-8400-9ff07b820e2b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 630.202416] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29698f75-e5b4-47ec-a660-9b710b1c1449 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.331818] env[63372]: DEBUG nova.compute.manager [req-717902e1-3178-40e9-9c29-00f7943b090e req-b99d1491-5754-4303-9fdb-7f0fcae9898c service nova] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Received event network-vif-deleted-61938f95-e588-4bdf-b61a-0795bd8279e3 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 630.551113] env[63372]: INFO nova.compute.manager [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] [instance: c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08] Took 1.03 seconds to deallocate network for instance. [ 630.563082] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023675, 'name': ReconfigVM_Task, 'duration_secs': 0.310853} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.563390] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Reconfigured VM instance instance-00000017 to attach disk [datastore1] 715408b8-0e76-4b61-a342-b168377cb288/715408b8-0e76-4b61-a342-b168377cb288.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 630.563981] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-56623961-ca51-4dd1-9d42-4128f1e37859 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.571137] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 630.571137] env[63372]: value = "task-1023676" [ 630.571137] env[63372]: _type = "Task" [ 630.571137] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.579525] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023676, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.666819] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Releasing lock "refresh_cache-ae86a977-0a9a-46e5-9072-cfb772324393" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.667296] env[63372]: DEBUG nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 630.667494] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 630.667891] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7a57b67a-0854-4246-a0ee-5ee733ee574b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.676664] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a502dad5-1e26-4911-9e1e-7856addd54ae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.700370] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ae86a977-0a9a-46e5-9072-cfb772324393 could not be found. [ 630.703089] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 630.703089] env[63372]: INFO nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Took 0.03 seconds to destroy the instance on the hypervisor. [ 630.703089] env[63372]: DEBUG oslo.service.loopingcall [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 630.703089] env[63372]: DEBUG nova.compute.manager [-] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 630.703089] env[63372]: DEBUG nova.network.neutron [-] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 630.707138] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 779e3559-37a5-49b2-b17a-bf95d0d8d23c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 630.722513] env[63372]: DEBUG nova.network.neutron [-] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.081608] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023676, 'name': Rename_Task, 'duration_secs': 0.172934} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.081888] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 631.082157] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad40c6fa-8d90-41a2-a84b-060379e9b46e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.089396] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 631.089396] env[63372]: value = "task-1023677" [ 631.089396] env[63372]: _type = "Task" [ 631.089396] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.098537] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023677, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.213297] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance cf673ac1-2c7d-468b-83ec-c723d5182457 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 631.226237] env[63372]: DEBUG nova.network.neutron [-] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.370562] env[63372]: DEBUG oslo_concurrency.lockutils [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Acquiring lock "8986e782-4d63-4491-8100-4341a3149812" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.374838] env[63372]: DEBUG oslo_concurrency.lockutils [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Lock "8986e782-4d63-4491-8100-4341a3149812" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.374838] env[63372]: DEBUG oslo_concurrency.lockutils [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Acquiring lock "8986e782-4d63-4491-8100-4341a3149812-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.374838] env[63372]: DEBUG oslo_concurrency.lockutils [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Lock "8986e782-4d63-4491-8100-4341a3149812-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.374838] env[63372]: DEBUG oslo_concurrency.lockutils [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Lock "8986e782-4d63-4491-8100-4341a3149812-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.375707] env[63372]: INFO nova.compute.manager [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Terminating instance [ 631.379249] env[63372]: DEBUG oslo_concurrency.lockutils [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Acquiring lock "refresh_cache-8986e782-4d63-4491-8100-4341a3149812" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.379249] env[63372]: DEBUG oslo_concurrency.lockutils [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Acquired lock "refresh_cache-8986e782-4d63-4491-8100-4341a3149812" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.379249] env[63372]: DEBUG nova.network.neutron [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 631.596996] env[63372]: INFO nova.scheduler.client.report [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Deleted allocations for instance c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08 [ 631.607065] env[63372]: DEBUG oslo_vmware.api [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023677, 'name': PowerOnVM_Task, 'duration_secs': 0.434862} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.608433] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 631.608748] env[63372]: DEBUG nova.compute.manager [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 631.610815] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b25e26-a30f-4134-aa60-23bc8531d688 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.715356] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance bbba9f28-045d-41ab-8539-5b2968fe3d54 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 631.731884] env[63372]: INFO nova.compute.manager [-] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Took 1.03 seconds to deallocate network for instance. [ 631.736191] env[63372]: DEBUG nova.compute.claims [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Aborting claim: {{(pid=63372) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 631.736468] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.900683] env[63372]: DEBUG nova.network.neutron [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.965893] env[63372]: DEBUG nova.network.neutron [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.109964] env[63372]: DEBUG oslo_concurrency.lockutils [None req-53e8eb66-aa69-4786-a46f-acb5cb2419cc tempest-FloatingIPsAssociationNegativeTestJSON-2140663510 tempest-FloatingIPsAssociationNegativeTestJSON-2140663510-project-member] Lock "c2fdcd5c-6d24-4ea7-8513-ae0d07ec4a08" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.301s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.125524] env[63372]: DEBUG oslo_concurrency.lockutils [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.220231] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 632.469371] env[63372]: DEBUG oslo_concurrency.lockutils [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Releasing lock "refresh_cache-8986e782-4d63-4491-8100-4341a3149812" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.469371] env[63372]: DEBUG nova.compute.manager [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 632.473017] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 632.473017] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b283833-e706-4414-8449-572b0bf6036b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.487152] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 632.487503] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf6d1871-e0b3-4631-af62-2fa8175f0456 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.496735] env[63372]: DEBUG oslo_vmware.api [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for the task: (returnval){ [ 632.496735] env[63372]: value = "task-1023678" [ 632.496735] env[63372]: _type = "Task" [ 632.496735] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.510055] env[63372]: DEBUG oslo_vmware.api [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023678, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.614008] env[63372]: DEBUG nova.compute.manager [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 632.724219] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 5adf6952-e019-4ea0-be91-8fe5fb83d53e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 633.007290] env[63372]: DEBUG oslo_vmware.api [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023678, 'name': PowerOffVM_Task, 'duration_secs': 0.132481} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.007918] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 633.007918] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 633.008055] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-62dd58e3-1c51-460a-883f-cafb7b4eaa2b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.035675] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 633.035735] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 633.035912] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Deleting the datastore file [datastore2] 8986e782-4d63-4491-8100-4341a3149812 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 633.036183] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94d79b9e-3164-4c27-b71e-bbc05f70de83 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.042624] env[63372]: DEBUG oslo_vmware.api [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for the task: (returnval){ [ 633.042624] env[63372]: value = "task-1023680" [ 633.042624] env[63372]: _type = "Task" [ 633.042624] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.050249] env[63372]: DEBUG oslo_vmware.api [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023680, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.149867] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.227751] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance ac90a156-be00-4f62-a76e-e08914531167 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 633.242100] env[63372]: INFO nova.compute.manager [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Rebuilding instance [ 633.296335] env[63372]: DEBUG nova.compute.manager [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 633.296335] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-294816e8-d91e-4d40-a93d-9d1355a35afb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.552343] env[63372]: DEBUG oslo_vmware.api [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Task: {'id': task-1023680, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095922} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.552765] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 633.555576] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 633.555576] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 633.555576] env[63372]: INFO nova.compute.manager [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] [instance: 8986e782-4d63-4491-8100-4341a3149812] Took 1.08 seconds to destroy the instance on the hypervisor. [ 633.555576] env[63372]: DEBUG oslo.service.loopingcall [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.555576] env[63372]: DEBUG nova.compute.manager [-] [instance: 8986e782-4d63-4491-8100-4341a3149812] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.555576] env[63372]: DEBUG nova.network.neutron [-] [instance: 8986e782-4d63-4491-8100-4341a3149812] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 633.571097] env[63372]: DEBUG nova.network.neutron [-] [instance: 8986e782-4d63-4491-8100-4341a3149812] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.732634] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 5819c38e-2cf9-4d16-b28a-5f23d35c3d44 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 633.811027] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 633.811027] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24a12dd0-58b3-470f-a7c1-220e43e26830 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.819126] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Waiting for the task: (returnval){ [ 633.819126] env[63372]: value = "task-1023681" [ 633.819126] env[63372]: _type = "Task" [ 633.819126] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.836054] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': task-1023681, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.871755] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Acquiring lock "f43555ef-b517-4b7d-9d2e-4787d40e201b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.872545] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Lock "f43555ef-b517-4b7d-9d2e-4787d40e201b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.075936] env[63372]: DEBUG nova.network.neutron [-] [instance: 8986e782-4d63-4491-8100-4341a3149812] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.243745] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance c4718797-aa86-4ec0-94d3-6480bd6aa898 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 634.337100] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': task-1023681, 'name': PowerOffVM_Task, 'duration_secs': 0.201126} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.337224] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 634.337364] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 634.338181] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19367e04-d16c-4fe2-8d6e-b0bc61d25642 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.347488] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 634.347682] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-29830507-549c-450f-8685-d1938483ddb8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.378634] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 634.378634] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 634.378846] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Deleting the datastore file [datastore1] 715408b8-0e76-4b61-a342-b168377cb288 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 634.378931] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b215c86-c6c0-4201-ba37-8a25ec88bd7e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.387867] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Waiting for the task: (returnval){ [ 634.387867] env[63372]: value = "task-1023683" [ 634.387867] env[63372]: _type = "Task" [ 634.387867] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.397139] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': task-1023683, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.581494] env[63372]: INFO nova.compute.manager [-] [instance: 8986e782-4d63-4491-8100-4341a3149812] Took 1.03 seconds to deallocate network for instance. [ 634.745931] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 634.896516] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': task-1023683, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096227} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.896839] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 634.897081] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 634.897278] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 635.090388] env[63372]: DEBUG oslo_concurrency.lockutils [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.248994] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 276c3490-240a-499b-ba64-4b02510b169c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 635.753111] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance d50919ac-3a0b-46ac-a837-ca3e6ad05173 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 635.945777] env[63372]: DEBUG nova.virt.hardware [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 635.945777] env[63372]: DEBUG nova.virt.hardware [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 635.946305] env[63372]: DEBUG nova.virt.hardware [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.946305] env[63372]: DEBUG nova.virt.hardware [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 635.946467] env[63372]: DEBUG nova.virt.hardware [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.946656] env[63372]: DEBUG nova.virt.hardware [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 635.946895] env[63372]: DEBUG nova.virt.hardware [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 635.947763] env[63372]: DEBUG nova.virt.hardware [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 635.947763] env[63372]: DEBUG nova.virt.hardware [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 635.947763] env[63372]: DEBUG nova.virt.hardware [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 635.947763] env[63372]: DEBUG nova.virt.hardware [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.948789] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb0f288-c653-44c9-8fcc-df8321d7be0c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.957778] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a94104-1287-4977-a424-df7986e12a4b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.971883] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Instance VIF info [] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 635.977658] env[63372]: DEBUG oslo.service.loopingcall [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 635.977974] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 635.978277] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c49ae489-f507-4e2f-966c-cf00783ccbae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.995249] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 635.995249] env[63372]: value = "task-1023684" [ 635.995249] env[63372]: _type = "Task" [ 635.995249] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.003114] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023684, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.256024] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 44fba254-7dec-4458-9ae3-fdbbe4895de5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 636.508112] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023684, 'name': CreateVM_Task, 'duration_secs': 0.257131} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.508400] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 636.508958] env[63372]: DEBUG oslo_concurrency.lockutils [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.509104] env[63372]: DEBUG oslo_concurrency.lockutils [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.509492] env[63372]: DEBUG oslo_concurrency.lockutils [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 636.509773] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85bb4bf5-aeff-43c6-a1de-a86c4df46cbb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.514293] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Waiting for the task: (returnval){ [ 636.514293] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52592853-9ca7-e303-12bf-679149b57486" [ 636.514293] env[63372]: _type = "Task" [ 636.514293] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.524166] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52592853-9ca7-e303-12bf-679149b57486, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.759799] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 85883c62-6d5e-4fa6-b4f0-afac984dc6bb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 637.024858] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52592853-9ca7-e303-12bf-679149b57486, 'name': SearchDatastore_Task, 'duration_secs': 0.008604} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.025732] env[63372]: DEBUG oslo_concurrency.lockutils [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.026045] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 637.026361] env[63372]: DEBUG oslo_concurrency.lockutils [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.026539] env[63372]: DEBUG oslo_concurrency.lockutils [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.026766] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 637.027080] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-78d6b9b9-be6f-4028-b6dc-fdc636c54b14 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.038720] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 637.038980] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 637.039783] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ceba24b7-375f-4986-9ab9-029c0a4f5532 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.045482] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Waiting for the task: (returnval){ [ 637.045482] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52519f4b-52ef-d555-b739-3582608ba3fa" [ 637.045482] env[63372]: _type = "Task" [ 637.045482] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.053188] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52519f4b-52ef-d555-b739-3582608ba3fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.263720] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance e5522005-8c54-43e1-ae23-5e5ff1ef0ee9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 637.556370] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52519f4b-52ef-d555-b739-3582608ba3fa, 'name': SearchDatastore_Task, 'duration_secs': 0.034111} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.557224] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1597a9b5-bfaa-4d8f-b13a-6b4167dd1443 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.563082] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Waiting for the task: (returnval){ [ 637.563082] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]526b825f-278d-da10-2a9b-8e649307181b" [ 637.563082] env[63372]: _type = "Task" [ 637.563082] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.571159] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526b825f-278d-da10-2a9b-8e649307181b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.768964] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance c825d2e2-a9cc-4c7a-b92e-039756d4121d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 638.084191] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526b825f-278d-da10-2a9b-8e649307181b, 'name': SearchDatastore_Task, 'duration_secs': 0.00966} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.084191] env[63372]: DEBUG oslo_concurrency.lockutils [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.084275] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 715408b8-0e76-4b61-a342-b168377cb288/715408b8-0e76-4b61-a342-b168377cb288.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 638.084568] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8350d82-2c9d-48c4-a33e-cbab936d1100 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.093121] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Waiting for the task: (returnval){ [ 638.093121] env[63372]: value = "task-1023685" [ 638.093121] env[63372]: _type = "Task" [ 638.093121] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.101571] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': task-1023685, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.275641] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 638.603170] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': task-1023685, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465884} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.603446] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 715408b8-0e76-4b61-a342-b168377cb288/715408b8-0e76-4b61-a342-b168377cb288.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 638.603657] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 638.603913] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a9180b33-a292-41cc-89e3-d64b6a16c524 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.610282] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Waiting for the task: (returnval){ [ 638.610282] env[63372]: value = "task-1023686" [ 638.610282] env[63372]: _type = "Task" [ 638.610282] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.621341] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': task-1023686, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.779543] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 63b9095c-fb49-4dc1-a6dc-96529aeaab81 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.123956] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': task-1023686, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061773} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.124314] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 639.125218] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3d1526-94e2-4b18-86e9-4633a3dc301a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.149986] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Reconfiguring VM instance instance-00000017 to attach disk [datastore2] 715408b8-0e76-4b61-a342-b168377cb288/715408b8-0e76-4b61-a342-b168377cb288.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 639.150777] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e5cf05b-bbd1-4352-836a-e8a5ec7af480 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.170895] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Waiting for the task: (returnval){ [ 639.170895] env[63372]: value = "task-1023687" [ 639.170895] env[63372]: _type = "Task" [ 639.170895] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.178962] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': task-1023687, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.282139] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance b182294d-2de8-4189-af7f-3e2d2c604a8b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.682406] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': task-1023687, 'name': ReconfigVM_Task, 'duration_secs': 0.292635} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.682939] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Reconfigured VM instance instance-00000017 to attach disk [datastore2] 715408b8-0e76-4b61-a342-b168377cb288/715408b8-0e76-4b61-a342-b168377cb288.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 639.683677] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5c490c4b-af7d-409d-8073-6b92cdc0a043 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.692422] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Waiting for the task: (returnval){ [ 639.692422] env[63372]: value = "task-1023688" [ 639.692422] env[63372]: _type = "Task" [ 639.692422] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.701608] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': task-1023688, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.786625] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance f0c60559-c072-4b61-afe8-03d6c131b307 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.204654] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': task-1023688, 'name': Rename_Task, 'duration_secs': 0.175167} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.204964] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 640.205260] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2aac8a53-7b3f-4211-8116-bdf112d65bcb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.211240] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Waiting for the task: (returnval){ [ 640.211240] env[63372]: value = "task-1023689" [ 640.211240] env[63372]: _type = "Task" [ 640.211240] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.219209] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': task-1023689, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.289013] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 8283b736-ad02-4082-97b7-561bd5c5da93 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.515244] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Acquiring lock "7a75a51d-b804-453b-ba7b-fda6c4931802" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.515548] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Lock "7a75a51d-b804-453b-ba7b-fda6c4931802" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.721051] env[63372]: DEBUG oslo_vmware.api [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Task: {'id': task-1023689, 'name': PowerOnVM_Task, 'duration_secs': 0.43134} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.721324] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 640.721521] env[63372]: DEBUG nova.compute.manager [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 640.722283] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dfa28a6-2544-48de-9df6-c6b7b2642351 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.792050] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 93a5d948-0629-4f53-a681-858d519acfa7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.792341] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 640.792484] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 641.210488] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a262bb-2e3c-407f-a76f-5dc4cf58c80f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.218744] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b77de89-709f-4acb-8011-b0a3b0217f11 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.254626] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c30fbc-7fd7-4831-838c-b1811c052b7d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.261640] env[63372]: DEBUG oslo_concurrency.lockutils [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.262709] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4efa445-aaf5-461b-b4d1-af6246c59f8d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.276333] env[63372]: DEBUG nova.compute.provider_tree [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.473116] env[63372]: DEBUG oslo_concurrency.lockutils [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquiring lock "715408b8-0e76-4b61-a342-b168377cb288" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.473388] env[63372]: DEBUG oslo_concurrency.lockutils [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Lock "715408b8-0e76-4b61-a342-b168377cb288" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.473598] env[63372]: DEBUG oslo_concurrency.lockutils [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquiring lock "715408b8-0e76-4b61-a342-b168377cb288-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.474460] env[63372]: DEBUG oslo_concurrency.lockutils [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Lock "715408b8-0e76-4b61-a342-b168377cb288-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.474460] env[63372]: DEBUG oslo_concurrency.lockutils [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Lock "715408b8-0e76-4b61-a342-b168377cb288-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.477623] env[63372]: INFO nova.compute.manager [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Terminating instance [ 641.482440] env[63372]: DEBUG oslo_concurrency.lockutils [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquiring lock "refresh_cache-715408b8-0e76-4b61-a342-b168377cb288" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.482598] env[63372]: DEBUG oslo_concurrency.lockutils [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquired lock "refresh_cache-715408b8-0e76-4b61-a342-b168377cb288" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.482757] env[63372]: DEBUG nova.network.neutron [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 641.641970] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "cd0c01ac-602b-44a3-8099-84b8a50b2449" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.642219] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "cd0c01ac-602b-44a3-8099-84b8a50b2449" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.779584] env[63372]: DEBUG nova.scheduler.client.report [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.004272] env[63372]: DEBUG nova.network.neutron [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.053913] env[63372]: DEBUG nova.network.neutron [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.285874] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 642.286148] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.133s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.286450] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.169s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.556932] env[63372]: DEBUG oslo_concurrency.lockutils [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Releasing lock "refresh_cache-715408b8-0e76-4b61-a342-b168377cb288" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.557375] env[63372]: DEBUG nova.compute.manager [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 642.557564] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 642.558468] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325c9990-a470-479d-a6ea-7f93368c3bb7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.566297] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 642.566536] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-796ea7af-102f-44ab-8d61-4421f9250f35 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.572876] env[63372]: DEBUG oslo_vmware.api [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 642.572876] env[63372]: value = "task-1023690" [ 642.572876] env[63372]: _type = "Task" [ 642.572876] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.580500] env[63372]: DEBUG oslo_vmware.api [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023690, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.084270] env[63372]: DEBUG oslo_vmware.api [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023690, 'name': PowerOffVM_Task, 'duration_secs': 0.173899} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.084537] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 643.084690] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 643.084932] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-89156844-6b3a-4fb4-9d89-1d3c15b9e699 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.105399] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac9a52b6-f34a-4839-ba87-0c49e479e4b3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.110673] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 643.110880] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 643.111127] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Deleting the datastore file [datastore2] 715408b8-0e76-4b61-a342-b168377cb288 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 643.111401] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-53e8dba7-6249-4666-8405-2820290f5e56 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.115988] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7311c69a-676e-4c29-bc40-b5a957cd4d36 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.120282] env[63372]: DEBUG oslo_vmware.api [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for the task: (returnval){ [ 643.120282] env[63372]: value = "task-1023692" [ 643.120282] env[63372]: _type = "Task" [ 643.120282] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.148968] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655ce2eb-e472-42c0-beaa-72d4e9888116 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.154381] env[63372]: DEBUG oslo_vmware.api [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023692, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.159288] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b83214-749e-4dcc-8284-15695fff54c5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.175711] env[63372]: DEBUG nova.compute.provider_tree [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.632403] env[63372]: DEBUG oslo_vmware.api [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Task: {'id': task-1023692, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.111531} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.632403] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 643.632403] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 643.632403] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 643.632403] env[63372]: INFO nova.compute.manager [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Took 1.07 seconds to destroy the instance on the hypervisor. [ 643.632586] env[63372]: DEBUG oslo.service.loopingcall [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.632586] env[63372]: DEBUG nova.compute.manager [-] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.632644] env[63372]: DEBUG nova.network.neutron [-] [instance: 715408b8-0e76-4b61-a342-b168377cb288] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 643.648845] env[63372]: DEBUG nova.network.neutron [-] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.679129] env[63372]: DEBUG nova.scheduler.client.report [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.152617] env[63372]: DEBUG nova.network.neutron [-] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.183981] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.897s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.184730] env[63372]: ERROR nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fb497c35-8c18-42a5-81e0-b59b655122c5, please check neutron logs for more information. [ 644.184730] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Traceback (most recent call last): [ 644.184730] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 644.184730] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] self.driver.spawn(context, instance, image_meta, [ 644.184730] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 644.184730] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.184730] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.184730] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] vm_ref = self.build_virtual_machine(instance, [ 644.184730] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.184730] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.184730] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.185036] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] for vif in network_info: [ 644.185036] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 644.185036] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] return self._sync_wrapper(fn, *args, **kwargs) [ 644.185036] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 644.185036] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] self.wait() [ 644.185036] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 644.185036] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] self[:] = self._gt.wait() [ 644.185036] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.185036] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] return self._exit_event.wait() [ 644.185036] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 644.185036] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] current.throw(*self._exc) [ 644.185036] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.185036] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] result = function(*args, **kwargs) [ 644.185368] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.185368] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] return func(*args, **kwargs) [ 644.185368] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.185368] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] raise e [ 644.185368] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.185368] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] nwinfo = self.network_api.allocate_for_instance( [ 644.185368] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.185368] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] created_port_ids = self._update_ports_for_instance( [ 644.185368] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.185368] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] with excutils.save_and_reraise_exception(): [ 644.185368] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.185368] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] self.force_reraise() [ 644.185368] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.185680] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] raise self.value [ 644.185680] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.185680] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] updated_port = self._update_port( [ 644.185680] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.185680] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] _ensure_no_port_binding_failure(port) [ 644.185680] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.185680] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] raise exception.PortBindingFailed(port_id=port['id']) [ 644.185680] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] nova.exception.PortBindingFailed: Binding failed for port fb497c35-8c18-42a5-81e0-b59b655122c5, please check neutron logs for more information. [ 644.185680] env[63372]: ERROR nova.compute.manager [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] [ 644.185680] env[63372]: DEBUG nova.compute.utils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Binding failed for port fb497c35-8c18-42a5-81e0-b59b655122c5, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 644.186691] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.964s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.189687] env[63372]: DEBUG nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Build of instance bd8cf6ba-8039-4c02-bda2-13c7d14f8903 was re-scheduled: Binding failed for port fb497c35-8c18-42a5-81e0-b59b655122c5, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 644.190123] env[63372]: DEBUG nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 644.190348] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "refresh_cache-bd8cf6ba-8039-4c02-bda2-13c7d14f8903" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.190488] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired lock "refresh_cache-bd8cf6ba-8039-4c02-bda2-13c7d14f8903" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.190641] env[63372]: DEBUG nova.network.neutron [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 644.658275] env[63372]: INFO nova.compute.manager [-] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Took 1.03 seconds to deallocate network for instance. [ 644.712505] env[63372]: DEBUG nova.network.neutron [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.804970] env[63372]: DEBUG nova.network.neutron [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.053815] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f918b7-9594-4a34-a1bd-e0012096f9f7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.061727] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fca5e83-511d-4f59-b286-4745feda2206 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.092989] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-456f7fd9-114e-412e-bd01-13932c946330 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.099913] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144e91b2-bb90-444b-b680-8492bdc71c53 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.112777] env[63372]: DEBUG nova.compute.provider_tree [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.165632] env[63372]: DEBUG oslo_concurrency.lockutils [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.307998] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Releasing lock "refresh_cache-bd8cf6ba-8039-4c02-bda2-13c7d14f8903" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.308311] env[63372]: DEBUG nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 645.308503] env[63372]: DEBUG nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.308674] env[63372]: DEBUG nova.network.neutron [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 645.323798] env[63372]: DEBUG nova.network.neutron [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.616105] env[63372]: DEBUG nova.scheduler.client.report [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.826907] env[63372]: DEBUG nova.network.neutron [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.121176] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.934s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.121775] env[63372]: ERROR nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 98f102f8-7a5b-4648-b3ce-a66979a361b5, please check neutron logs for more information. [ 646.121775] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Traceback (most recent call last): [ 646.121775] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 646.121775] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] self.driver.spawn(context, instance, image_meta, [ 646.121775] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 646.121775] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.121775] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.121775] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] vm_ref = self.build_virtual_machine(instance, [ 646.121775] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.121775] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.121775] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.122073] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] for vif in network_info: [ 646.122073] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 646.122073] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] return self._sync_wrapper(fn, *args, **kwargs) [ 646.122073] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 646.122073] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] self.wait() [ 646.122073] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 646.122073] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] self[:] = self._gt.wait() [ 646.122073] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.122073] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] return self._exit_event.wait() [ 646.122073] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 646.122073] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] current.throw(*self._exc) [ 646.122073] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.122073] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] result = function(*args, **kwargs) [ 646.122373] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 646.122373] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] return func(*args, **kwargs) [ 646.122373] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.122373] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] raise e [ 646.122373] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.122373] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] nwinfo = self.network_api.allocate_for_instance( [ 646.122373] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 646.122373] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] created_port_ids = self._update_ports_for_instance( [ 646.122373] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 646.122373] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] with excutils.save_and_reraise_exception(): [ 646.122373] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.122373] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] self.force_reraise() [ 646.122373] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.122668] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] raise self.value [ 646.122668] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 646.122668] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] updated_port = self._update_port( [ 646.122668] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.122668] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] _ensure_no_port_binding_failure(port) [ 646.122668] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.122668] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] raise exception.PortBindingFailed(port_id=port['id']) [ 646.122668] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] nova.exception.PortBindingFailed: Binding failed for port 98f102f8-7a5b-4648-b3ce-a66979a361b5, please check neutron logs for more information. [ 646.122668] env[63372]: ERROR nova.compute.manager [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] [ 646.122668] env[63372]: DEBUG nova.compute.utils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Binding failed for port 98f102f8-7a5b-4648-b3ce-a66979a361b5, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 646.123732] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.382s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.125169] env[63372]: INFO nova.compute.claims [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 646.127878] env[63372]: DEBUG nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Build of instance 6aa468f8-cd63-459b-8302-8bc22580b01d was re-scheduled: Binding failed for port 98f102f8-7a5b-4648-b3ce-a66979a361b5, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 646.130050] env[63372]: DEBUG nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 646.130050] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Acquiring lock "refresh_cache-6aa468f8-cd63-459b-8302-8bc22580b01d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.130050] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Acquired lock "refresh_cache-6aa468f8-cd63-459b-8302-8bc22580b01d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.130050] env[63372]: DEBUG nova.network.neutron [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 646.329803] env[63372]: INFO nova.compute.manager [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: bd8cf6ba-8039-4c02-bda2-13c7d14f8903] Took 1.02 seconds to deallocate network for instance. [ 646.648176] env[63372]: DEBUG nova.network.neutron [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 646.716884] env[63372]: DEBUG nova.network.neutron [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.222090] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Releasing lock "refresh_cache-6aa468f8-cd63-459b-8302-8bc22580b01d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.222883] env[63372]: DEBUG nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 647.222883] env[63372]: DEBUG nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 647.222883] env[63372]: DEBUG nova.network.neutron [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 647.238749] env[63372]: DEBUG nova.network.neutron [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.362909] env[63372]: INFO nova.scheduler.client.report [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Deleted allocations for instance bd8cf6ba-8039-4c02-bda2-13c7d14f8903 [ 647.465976] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa1546c-fb0e-45e8-9a1e-b92f2066d73f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.474298] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b03137-eb6f-49bd-a868-da1c5b4a5c82 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.504125] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25bef89-19c6-4909-aa38-aa6a01f358d0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.511648] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc42b3cf-962a-4414-bda6-70f0cad30501 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.524695] env[63372]: DEBUG nova.compute.provider_tree [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.741864] env[63372]: DEBUG nova.network.neutron [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.874068] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2f0308d-ac4e-4438-8cdb-d42a16663a05 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "bd8cf6ba-8039-4c02-bda2-13c7d14f8903" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.257s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.027985] env[63372]: DEBUG nova.scheduler.client.report [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 648.244995] env[63372]: INFO nova.compute.manager [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] [instance: 6aa468f8-cd63-459b-8302-8bc22580b01d] Took 1.02 seconds to deallocate network for instance. [ 648.376991] env[63372]: DEBUG nova.compute.manager [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 648.532521] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.533223] env[63372]: DEBUG nova.compute.manager [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 648.538169] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.771s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.540304] env[63372]: INFO nova.compute.claims [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.896540] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.044592] env[63372]: DEBUG nova.compute.utils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 649.047608] env[63372]: DEBUG nova.compute.manager [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 649.047786] env[63372]: DEBUG nova.network.neutron [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 649.090571] env[63372]: DEBUG nova.policy [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2124bf74e1944c7ac4e22dc5fea8bc3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '056172ef521c4de08b5033057773b2ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 649.156673] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "ae6f3f42-7213-4ab1-b74c-1a557df6748b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.156673] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "ae6f3f42-7213-4ab1-b74c-1a557df6748b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.278621] env[63372]: INFO nova.scheduler.client.report [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Deleted allocations for instance 6aa468f8-cd63-459b-8302-8bc22580b01d [ 649.374014] env[63372]: DEBUG nova.network.neutron [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Successfully created port: 7f75cf3b-acb2-456d-896e-25184e6811c4 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 649.551727] env[63372]: DEBUG nova.compute.manager [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 649.613765] env[63372]: DEBUG nova.network.neutron [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Successfully created port: d1e8b347-53e9-4aeb-82d6-f9e255d24442 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 649.791249] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a156296-cfdb-4ef2-98aa-51e62ea3fb39 tempest-ServerMetadataTestJSON-1605894566 tempest-ServerMetadataTestJSON-1605894566-project-member] Lock "6aa468f8-cd63-459b-8302-8bc22580b01d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.185s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.912016] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad07c108-965f-4c8b-9216-f541f9d314f7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.918278] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e212b3-5601-45f7-a762-4b61b77ee9b1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.957361] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c513f3e-e910-47ac-b82a-fda422b9458e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.966252] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3574d155-419f-45c7-8318-87d2ef9503b6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.981141] env[63372]: DEBUG nova.compute.provider_tree [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.015486] env[63372]: DEBUG nova.network.neutron [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Successfully created port: a52a68b2-6f6c-4b4d-b377-b79e60f46dac {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 650.294034] env[63372]: DEBUG nova.compute.manager [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 650.483912] env[63372]: DEBUG nova.scheduler.client.report [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.563283] env[63372]: DEBUG nova.compute.manager [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 650.589947] env[63372]: DEBUG nova.virt.hardware [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 650.590438] env[63372]: DEBUG nova.virt.hardware [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 650.590727] env[63372]: DEBUG nova.virt.hardware [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 650.591063] env[63372]: DEBUG nova.virt.hardware [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 650.592481] env[63372]: DEBUG nova.virt.hardware [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 650.592481] env[63372]: DEBUG nova.virt.hardware [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 650.592481] env[63372]: DEBUG nova.virt.hardware [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 650.592481] env[63372]: DEBUG nova.virt.hardware [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 650.592481] env[63372]: DEBUG nova.virt.hardware [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 650.592705] env[63372]: DEBUG nova.virt.hardware [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 650.592705] env[63372]: DEBUG nova.virt.hardware [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 650.593470] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9019e902-5b02-41e0-abb6-6c8a60757aaf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.601872] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0159da21-63a0-421f-af58-f46db4195a08 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.816277] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.989679] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.990227] env[63372]: DEBUG nova.compute.manager [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 650.992833] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.801s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.997038] env[63372]: INFO nova.compute.claims [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 651.498848] env[63372]: DEBUG nova.compute.utils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 651.502497] env[63372]: DEBUG nova.compute.manager [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 651.502669] env[63372]: DEBUG nova.network.neutron [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 651.589074] env[63372]: DEBUG nova.compute.manager [req-2031cd49-817d-4c06-9501-531d7ca5c055 req-12cbee2e-5931-47d3-9d00-8ed1ee8b60d6 service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Received event network-vif-plugged-7f75cf3b-acb2-456d-896e-25184e6811c4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.589331] env[63372]: DEBUG oslo_concurrency.lockutils [req-2031cd49-817d-4c06-9501-531d7ca5c055 req-12cbee2e-5931-47d3-9d00-8ed1ee8b60d6 service nova] Acquiring lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.589517] env[63372]: DEBUG oslo_concurrency.lockutils [req-2031cd49-817d-4c06-9501-531d7ca5c055 req-12cbee2e-5931-47d3-9d00-8ed1ee8b60d6 service nova] Lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.589679] env[63372]: DEBUG oslo_concurrency.lockutils [req-2031cd49-817d-4c06-9501-531d7ca5c055 req-12cbee2e-5931-47d3-9d00-8ed1ee8b60d6 service nova] Lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.589842] env[63372]: DEBUG nova.compute.manager [req-2031cd49-817d-4c06-9501-531d7ca5c055 req-12cbee2e-5931-47d3-9d00-8ed1ee8b60d6 service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] No waiting events found dispatching network-vif-plugged-7f75cf3b-acb2-456d-896e-25184e6811c4 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 651.590009] env[63372]: WARNING nova.compute.manager [req-2031cd49-817d-4c06-9501-531d7ca5c055 req-12cbee2e-5931-47d3-9d00-8ed1ee8b60d6 service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Received unexpected event network-vif-plugged-7f75cf3b-acb2-456d-896e-25184e6811c4 for instance with vm_state building and task_state spawning. [ 651.599423] env[63372]: DEBUG nova.policy [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34bb855019714168bd59140697834c99', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88c54a0f2cd941d6b0ef1ba5483dfb59', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 651.855792] env[63372]: DEBUG nova.network.neutron [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Successfully updated port: 7f75cf3b-acb2-456d-896e-25184e6811c4 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 652.003691] env[63372]: DEBUG nova.compute.manager [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 652.015699] env[63372]: DEBUG nova.network.neutron [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Successfully created port: 8815dc60-cd0a-42fb-8e16-875baf0e4298 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 652.450328] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-198db450-4de5-4f49-ab2b-e40c4c844199 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.459080] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f014b54f-3f2e-4fea-af45-da857ec90b20 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.490554] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7956c862-88bd-43db-9f7d-7d2905dc9009 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.499179] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a363b2-a42f-4850-92d3-4a9992bf2422 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.516868] env[63372]: DEBUG nova.compute.provider_tree [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.019770] env[63372]: DEBUG nova.compute.manager [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 653.022468] env[63372]: DEBUG nova.scheduler.client.report [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 653.050071] env[63372]: DEBUG nova.virt.hardware [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 653.050545] env[63372]: DEBUG nova.virt.hardware [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 653.050815] env[63372]: DEBUG nova.virt.hardware [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 653.051128] env[63372]: DEBUG nova.virt.hardware [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 653.055017] env[63372]: DEBUG nova.virt.hardware [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 653.055017] env[63372]: DEBUG nova.virt.hardware [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 653.055017] env[63372]: DEBUG nova.virt.hardware [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 653.055017] env[63372]: DEBUG nova.virt.hardware [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 653.055017] env[63372]: DEBUG nova.virt.hardware [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 653.055232] env[63372]: DEBUG nova.virt.hardware [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 653.055232] env[63372]: DEBUG nova.virt.hardware [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 653.055232] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a69791-f11d-4674-b781-c355de4c751e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.062606] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66871b92-33ae-45ea-8f8c-73e0dd1735e7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.527151] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.527913] env[63372]: DEBUG nova.compute.manager [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 653.530420] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.335s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.531991] env[63372]: INFO nova.compute.claims [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 653.616341] env[63372]: DEBUG nova.compute.manager [req-d2a7f41b-362c-4e70-b752-e4dd8683e754 req-dc15c2b7-e318-4b15-8865-298b9e9cb329 service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Received event network-changed-7f75cf3b-acb2-456d-896e-25184e6811c4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 653.616515] env[63372]: DEBUG nova.compute.manager [req-d2a7f41b-362c-4e70-b752-e4dd8683e754 req-dc15c2b7-e318-4b15-8865-298b9e9cb329 service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Refreshing instance network info cache due to event network-changed-7f75cf3b-acb2-456d-896e-25184e6811c4. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 653.616720] env[63372]: DEBUG oslo_concurrency.lockutils [req-d2a7f41b-362c-4e70-b752-e4dd8683e754 req-dc15c2b7-e318-4b15-8865-298b9e9cb329 service nova] Acquiring lock "refresh_cache-6d4502b7-b9df-4fef-8400-9ff07b820e2b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.616860] env[63372]: DEBUG oslo_concurrency.lockutils [req-d2a7f41b-362c-4e70-b752-e4dd8683e754 req-dc15c2b7-e318-4b15-8865-298b9e9cb329 service nova] Acquired lock "refresh_cache-6d4502b7-b9df-4fef-8400-9ff07b820e2b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.617144] env[63372]: DEBUG nova.network.neutron [req-d2a7f41b-362c-4e70-b752-e4dd8683e754 req-dc15c2b7-e318-4b15-8865-298b9e9cb329 service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Refreshing network info cache for port 7f75cf3b-acb2-456d-896e-25184e6811c4 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 653.904891] env[63372]: DEBUG nova.network.neutron [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Successfully updated port: 8815dc60-cd0a-42fb-8e16-875baf0e4298 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 654.037633] env[63372]: DEBUG nova.compute.utils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 654.043943] env[63372]: DEBUG nova.compute.manager [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 654.043943] env[63372]: DEBUG nova.network.neutron [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 654.082925] env[63372]: DEBUG nova.network.neutron [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Successfully updated port: d1e8b347-53e9-4aeb-82d6-f9e255d24442 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 654.103797] env[63372]: DEBUG nova.policy [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d57dfd73d0f4642bd2b7dd2a4d989cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06cecd1557de407697ad403a85138481', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 654.174276] env[63372]: DEBUG nova.network.neutron [req-d2a7f41b-362c-4e70-b752-e4dd8683e754 req-dc15c2b7-e318-4b15-8865-298b9e9cb329 service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.326604] env[63372]: DEBUG nova.network.neutron [req-d2a7f41b-362c-4e70-b752-e4dd8683e754 req-dc15c2b7-e318-4b15-8865-298b9e9cb329 service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.409680] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Acquiring lock "refresh_cache-779e3559-37a5-49b2-b17a-bf95d0d8d23c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.409846] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Acquired lock "refresh_cache-779e3559-37a5-49b2-b17a-bf95d0d8d23c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.410011] env[63372]: DEBUG nova.network.neutron [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 654.499928] env[63372]: DEBUG nova.network.neutron [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Successfully created port: e56add8e-4342-4ab4-b2f2-c03f257f3d3a {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 654.546763] env[63372]: DEBUG nova.compute.manager [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 654.829690] env[63372]: DEBUG oslo_concurrency.lockutils [req-d2a7f41b-362c-4e70-b752-e4dd8683e754 req-dc15c2b7-e318-4b15-8865-298b9e9cb329 service nova] Releasing lock "refresh_cache-6d4502b7-b9df-4fef-8400-9ff07b820e2b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.942707] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0670be1f-0202-464d-bdce-b442fa1e52fd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.950901] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36aba2e7-5564-4746-b03e-c46bb75d3a38 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.989019] env[63372]: DEBUG nova.network.neutron [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.990678] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb3aad4-0da4-4a46-a671-4d56f79d4081 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.998924] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecbaf473-927d-4e87-9cb1-3f51782ddf67 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.012594] env[63372]: DEBUG nova.compute.provider_tree [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.247989] env[63372]: DEBUG nova.network.neutron [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Updating instance_info_cache with network_info: [{"id": "8815dc60-cd0a-42fb-8e16-875baf0e4298", "address": "fa:16:3e:26:42:dd", "network": {"id": "bc9a78c9-a249-4eed-8ec4-bb2a330b76c3", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1602987910-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88c54a0f2cd941d6b0ef1ba5483dfb59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604c9724-b4ef-4393-a76e-eb4a2b510796", "external-id": "nsx-vlan-transportzone-909", "segmentation_id": 909, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8815dc60-cd", "ovs_interfaceid": "8815dc60-cd0a-42fb-8e16-875baf0e4298", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.515565] env[63372]: DEBUG nova.scheduler.client.report [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.558724] env[63372]: DEBUG nova.compute.manager [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 655.584905] env[63372]: DEBUG nova.virt.hardware [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 655.585167] env[63372]: DEBUG nova.virt.hardware [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 655.585322] env[63372]: DEBUG nova.virt.hardware [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 655.585499] env[63372]: DEBUG nova.virt.hardware [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 655.585674] env[63372]: DEBUG nova.virt.hardware [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 655.585842] env[63372]: DEBUG nova.virt.hardware [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 655.586030] env[63372]: DEBUG nova.virt.hardware [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 655.586197] env[63372]: DEBUG nova.virt.hardware [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 655.586361] env[63372]: DEBUG nova.virt.hardware [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 655.586524] env[63372]: DEBUG nova.virt.hardware [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 655.586738] env[63372]: DEBUG nova.virt.hardware [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 655.587620] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da05a4d2-fc65-4437-b4e5-485677cf5872 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.595781] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e4e33d-8510-4a30-95dc-8104ff795867 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.681435] env[63372]: DEBUG nova.compute.manager [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Received event network-vif-plugged-8815dc60-cd0a-42fb-8e16-875baf0e4298 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 655.681435] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] Acquiring lock "779e3559-37a5-49b2-b17a-bf95d0d8d23c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.681435] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] Lock "779e3559-37a5-49b2-b17a-bf95d0d8d23c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.682096] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] Lock "779e3559-37a5-49b2-b17a-bf95d0d8d23c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.682424] env[63372]: DEBUG nova.compute.manager [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] No waiting events found dispatching network-vif-plugged-8815dc60-cd0a-42fb-8e16-875baf0e4298 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 655.682722] env[63372]: WARNING nova.compute.manager [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Received unexpected event network-vif-plugged-8815dc60-cd0a-42fb-8e16-875baf0e4298 for instance with vm_state building and task_state spawning. [ 655.683016] env[63372]: DEBUG nova.compute.manager [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Received event network-changed-8815dc60-cd0a-42fb-8e16-875baf0e4298 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 655.683382] env[63372]: DEBUG nova.compute.manager [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Refreshing instance network info cache due to event network-changed-8815dc60-cd0a-42fb-8e16-875baf0e4298. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 655.683691] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] Acquiring lock "refresh_cache-779e3559-37a5-49b2-b17a-bf95d0d8d23c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.752497] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Releasing lock "refresh_cache-779e3559-37a5-49b2-b17a-bf95d0d8d23c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.752497] env[63372]: DEBUG nova.compute.manager [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Instance network_info: |[{"id": "8815dc60-cd0a-42fb-8e16-875baf0e4298", "address": "fa:16:3e:26:42:dd", "network": {"id": "bc9a78c9-a249-4eed-8ec4-bb2a330b76c3", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1602987910-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88c54a0f2cd941d6b0ef1ba5483dfb59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604c9724-b4ef-4393-a76e-eb4a2b510796", "external-id": "nsx-vlan-transportzone-909", "segmentation_id": 909, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8815dc60-cd", "ovs_interfaceid": "8815dc60-cd0a-42fb-8e16-875baf0e4298", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 655.752693] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] Acquired lock "refresh_cache-779e3559-37a5-49b2-b17a-bf95d0d8d23c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.752693] env[63372]: DEBUG nova.network.neutron [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Refreshing network info cache for port 8815dc60-cd0a-42fb-8e16-875baf0e4298 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 655.753991] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:42:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '604c9724-b4ef-4393-a76e-eb4a2b510796', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8815dc60-cd0a-42fb-8e16-875baf0e4298', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 655.768229] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Creating folder: Project (88c54a0f2cd941d6b0ef1ba5483dfb59). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 655.769210] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a0b1f239-5bad-42af-b8b7-631faa8237cd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.780847] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Created folder: Project (88c54a0f2cd941d6b0ef1ba5483dfb59) in parent group-v227230. [ 655.783019] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Creating folder: Instances. Parent ref: group-v227243. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 655.783019] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6a0353c4-2b27-4be1-ae2d-aa4003c35782 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.790497] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Created folder: Instances in parent group-v227243. [ 655.790851] env[63372]: DEBUG oslo.service.loopingcall [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.791157] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 655.791661] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ab67d80b-9d89-4d6c-891a-78f0186cf205 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.814047] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 655.814047] env[63372]: value = "task-1023695" [ 655.814047] env[63372]: _type = "Task" [ 655.814047] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.821849] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023695, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.023024] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.023663] env[63372]: DEBUG nova.compute.manager [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 656.030845] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.794s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.032379] env[63372]: INFO nova.compute.claims [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 656.247359] env[63372]: DEBUG nova.network.neutron [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Successfully updated port: e56add8e-4342-4ab4-b2f2-c03f257f3d3a {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 656.324284] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023695, 'name': CreateVM_Task, 'duration_secs': 0.31121} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.325135] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 656.337531] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.337712] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.338083] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 656.338295] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c633252-f021-4dff-80d6-bf5e6aaa9ca7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.343359] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Waiting for the task: (returnval){ [ 656.343359] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52d41a14-2571-5615-a823-ff25b893b5ee" [ 656.343359] env[63372]: _type = "Task" [ 656.343359] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.351408] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52d41a14-2571-5615-a823-ff25b893b5ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.402091] env[63372]: DEBUG nova.network.neutron [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Successfully updated port: a52a68b2-6f6c-4b4d-b377-b79e60f46dac {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 656.537020] env[63372]: DEBUG nova.compute.utils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 656.541229] env[63372]: DEBUG nova.compute.manager [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 656.541524] env[63372]: DEBUG nova.network.neutron [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 656.563656] env[63372]: DEBUG nova.network.neutron [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Updated VIF entry in instance network info cache for port 8815dc60-cd0a-42fb-8e16-875baf0e4298. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 656.564016] env[63372]: DEBUG nova.network.neutron [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Updating instance_info_cache with network_info: [{"id": "8815dc60-cd0a-42fb-8e16-875baf0e4298", "address": "fa:16:3e:26:42:dd", "network": {"id": "bc9a78c9-a249-4eed-8ec4-bb2a330b76c3", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1602987910-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88c54a0f2cd941d6b0ef1ba5483dfb59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604c9724-b4ef-4393-a76e-eb4a2b510796", "external-id": "nsx-vlan-transportzone-909", "segmentation_id": 909, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8815dc60-cd", "ovs_interfaceid": "8815dc60-cd0a-42fb-8e16-875baf0e4298", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.587035] env[63372]: DEBUG nova.policy [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7ffe6776b664c00b9086046f9171ec6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46d2cb188a004dc385f2aa15458be929', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 656.751625] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Acquiring lock "refresh_cache-cf673ac1-2c7d-468b-83ec-c723d5182457" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.751784] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Acquired lock "refresh_cache-cf673ac1-2c7d-468b-83ec-c723d5182457" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.751933] env[63372]: DEBUG nova.network.neutron [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 656.857056] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52d41a14-2571-5615-a823-ff25b893b5ee, 'name': SearchDatastore_Task, 'duration_secs': 0.011651} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.857713] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.857958] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 656.858209] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.858356] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.858533] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 656.858849] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-92adeb63-6164-4b1d-932b-a440f4dd42fa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.866698] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 656.867120] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 656.868217] env[63372]: DEBUG nova.network.neutron [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Successfully created port: b4cf27a3-fc17-4ac5-b96a-3be058f71e03 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 656.870064] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29b3fb7b-9ab6-4cc6-a9cf-e54a4944fb9a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.875554] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Waiting for the task: (returnval){ [ 656.875554] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52ee0b1c-ba15-9dd6-8e16-2ded053510a2" [ 656.875554] env[63372]: _type = "Task" [ 656.875554] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.883435] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ee0b1c-ba15-9dd6-8e16-2ded053510a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.905808] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "refresh_cache-6d4502b7-b9df-4fef-8400-9ff07b820e2b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.905955] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquired lock "refresh_cache-6d4502b7-b9df-4fef-8400-9ff07b820e2b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.906121] env[63372]: DEBUG nova.network.neutron [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 657.041102] env[63372]: DEBUG nova.compute.manager [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 657.071375] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] Releasing lock "refresh_cache-779e3559-37a5-49b2-b17a-bf95d0d8d23c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.071375] env[63372]: DEBUG nova.compute.manager [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Received event network-vif-plugged-d1e8b347-53e9-4aeb-82d6-f9e255d24442 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 657.071375] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] Acquiring lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.071375] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] Lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.071375] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] Lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.071672] env[63372]: DEBUG nova.compute.manager [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] No waiting events found dispatching network-vif-plugged-d1e8b347-53e9-4aeb-82d6-f9e255d24442 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 657.071672] env[63372]: WARNING nova.compute.manager [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Received unexpected event network-vif-plugged-d1e8b347-53e9-4aeb-82d6-f9e255d24442 for instance with vm_state building and task_state spawning. [ 657.071672] env[63372]: DEBUG nova.compute.manager [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Received event network-changed-d1e8b347-53e9-4aeb-82d6-f9e255d24442 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 657.071672] env[63372]: DEBUG nova.compute.manager [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Refreshing instance network info cache due to event network-changed-d1e8b347-53e9-4aeb-82d6-f9e255d24442. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 657.071672] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] Acquiring lock "refresh_cache-6d4502b7-b9df-4fef-8400-9ff07b820e2b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.283315] env[63372]: DEBUG nova.network.neutron [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.384893] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ee0b1c-ba15-9dd6-8e16-2ded053510a2, 'name': SearchDatastore_Task, 'duration_secs': 0.007991} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.388211] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10acd310-4136-4bb4-b2a3-0b057e22af84 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.392948] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Waiting for the task: (returnval){ [ 657.392948] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52609b32-ce6c-92e1-1151-3289f4fff4bc" [ 657.392948] env[63372]: _type = "Task" [ 657.392948] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.405874] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52609b32-ce6c-92e1-1151-3289f4fff4bc, 'name': SearchDatastore_Task, 'duration_secs': 0.008737} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.405874] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.405874] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 779e3559-37a5-49b2-b17a-bf95d0d8d23c/779e3559-37a5-49b2-b17a-bf95d0d8d23c.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 657.405874] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f5a81382-896f-4c47-8258-3651b5375882 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.411614] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8754749-4d6c-4035-ba89-2c811f8d8872 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.415864] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Waiting for the task: (returnval){ [ 657.415864] env[63372]: value = "task-1023696" [ 657.415864] env[63372]: _type = "Task" [ 657.415864] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.421763] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-939f4eab-cb6b-4baf-9559-c0676f6f9124 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.427563] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': task-1023696, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.453861] env[63372]: DEBUG nova.network.neutron [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Updating instance_info_cache with network_info: [{"id": "e56add8e-4342-4ab4-b2f2-c03f257f3d3a", "address": "fa:16:3e:11:78:8c", "network": {"id": "79050166-5dc7-4ef3-9819-ac534e041985", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-2103555585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06cecd1557de407697ad403a85138481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24727047-6358-4015-86c1-394ab07fb88f", "external-id": "nsx-vlan-transportzone-476", "segmentation_id": 476, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape56add8e-43", "ovs_interfaceid": "e56add8e-4342-4ab4-b2f2-c03f257f3d3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.455595] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e998dd7-8ed9-43dd-bd59-c0e5c0bda9af {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.463129] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bebd4d9f-34a8-4827-8ad3-26ffa83e5b93 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.476960] env[63372]: DEBUG nova.compute.provider_tree [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.478614] env[63372]: DEBUG nova.network.neutron [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.885997] env[63372]: DEBUG nova.compute.manager [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Received event network-vif-plugged-e56add8e-4342-4ab4-b2f2-c03f257f3d3a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 657.886231] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] Acquiring lock "cf673ac1-2c7d-468b-83ec-c723d5182457-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.886433] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] Lock "cf673ac1-2c7d-468b-83ec-c723d5182457-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.886615] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] Lock "cf673ac1-2c7d-468b-83ec-c723d5182457-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.886789] env[63372]: DEBUG nova.compute.manager [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] No waiting events found dispatching network-vif-plugged-e56add8e-4342-4ab4-b2f2-c03f257f3d3a {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 657.886949] env[63372]: WARNING nova.compute.manager [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Received unexpected event network-vif-plugged-e56add8e-4342-4ab4-b2f2-c03f257f3d3a for instance with vm_state building and task_state spawning. [ 657.887113] env[63372]: DEBUG nova.compute.manager [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Received event network-vif-plugged-a52a68b2-6f6c-4b4d-b377-b79e60f46dac {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 657.887267] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] Acquiring lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.887436] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] Lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.887581] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] Lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.887732] env[63372]: DEBUG nova.compute.manager [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] No waiting events found dispatching network-vif-plugged-a52a68b2-6f6c-4b4d-b377-b79e60f46dac {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 657.887934] env[63372]: WARNING nova.compute.manager [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Received unexpected event network-vif-plugged-a52a68b2-6f6c-4b4d-b377-b79e60f46dac for instance with vm_state building and task_state spawning. [ 657.888147] env[63372]: DEBUG nova.compute.manager [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Received event network-changed-e56add8e-4342-4ab4-b2f2-c03f257f3d3a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 657.888265] env[63372]: DEBUG nova.compute.manager [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Refreshing instance network info cache due to event network-changed-e56add8e-4342-4ab4-b2f2-c03f257f3d3a. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 657.888421] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] Acquiring lock "refresh_cache-cf673ac1-2c7d-468b-83ec-c723d5182457" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.926468] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': task-1023696, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443677} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.926708] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 779e3559-37a5-49b2-b17a-bf95d0d8d23c/779e3559-37a5-49b2-b17a-bf95d0d8d23c.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 657.926910] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 657.927155] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b6c32130-cf40-4b8b-972b-80c7326816f3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.933723] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Waiting for the task: (returnval){ [ 657.933723] env[63372]: value = "task-1023697" [ 657.933723] env[63372]: _type = "Task" [ 657.933723] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.940998] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': task-1023697, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.960444] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Releasing lock "refresh_cache-cf673ac1-2c7d-468b-83ec-c723d5182457" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.960754] env[63372]: DEBUG nova.compute.manager [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Instance network_info: |[{"id": "e56add8e-4342-4ab4-b2f2-c03f257f3d3a", "address": "fa:16:3e:11:78:8c", "network": {"id": "79050166-5dc7-4ef3-9819-ac534e041985", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-2103555585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06cecd1557de407697ad403a85138481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24727047-6358-4015-86c1-394ab07fb88f", "external-id": "nsx-vlan-transportzone-476", "segmentation_id": 476, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape56add8e-43", "ovs_interfaceid": "e56add8e-4342-4ab4-b2f2-c03f257f3d3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 657.961423] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] Acquired lock "refresh_cache-cf673ac1-2c7d-468b-83ec-c723d5182457" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.961652] env[63372]: DEBUG nova.network.neutron [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Refreshing network info cache for port e56add8e-4342-4ab4-b2f2-c03f257f3d3a {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 657.962830] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:78:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24727047-6358-4015-86c1-394ab07fb88f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e56add8e-4342-4ab4-b2f2-c03f257f3d3a', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 657.971468] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Creating folder: Project (06cecd1557de407697ad403a85138481). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 657.971980] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-228ae65d-953d-46e9-b351-569687aa7698 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.981600] env[63372]: DEBUG nova.scheduler.client.report [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.986248] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Created folder: Project (06cecd1557de407697ad403a85138481) in parent group-v227230. [ 657.986427] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Creating folder: Instances. Parent ref: group-v227246. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 657.986651] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b49d8aa4-21fc-40d8-b878-d30c09b34072 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.996131] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Created folder: Instances in parent group-v227246. [ 657.996131] env[63372]: DEBUG oslo.service.loopingcall [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 657.996131] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 657.996257] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-240f0a92-5686-49e9-a0e0-a9c7c479c5d7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.020232] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 658.020232] env[63372]: value = "task-1023700" [ 658.020232] env[63372]: _type = "Task" [ 658.020232] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.029014] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023700, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.051689] env[63372]: DEBUG nova.compute.manager [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 658.092823] env[63372]: DEBUG nova.virt.hardware [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 658.093048] env[63372]: DEBUG nova.virt.hardware [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 658.093219] env[63372]: DEBUG nova.virt.hardware [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 658.093402] env[63372]: DEBUG nova.virt.hardware [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 658.093578] env[63372]: DEBUG nova.virt.hardware [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 658.093737] env[63372]: DEBUG nova.virt.hardware [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 658.093938] env[63372]: DEBUG nova.virt.hardware [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 658.094106] env[63372]: DEBUG nova.virt.hardware [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 658.094295] env[63372]: DEBUG nova.virt.hardware [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 658.094422] env[63372]: DEBUG nova.virt.hardware [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 658.094594] env[63372]: DEBUG nova.virt.hardware [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 658.095762] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4fa1865-41c3-48ac-bb10-4ef6be559384 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.107033] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222230a3-bd3f-43ca-9631-505c6500d974 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.391487] env[63372]: DEBUG nova.network.neutron [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Successfully updated port: b4cf27a3-fc17-4ac5-b96a-3be058f71e03 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 658.429658] env[63372]: DEBUG nova.network.neutron [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Updating instance_info_cache with network_info: [{"id": "7f75cf3b-acb2-456d-896e-25184e6811c4", "address": "fa:16:3e:f0:dc:5b", "network": {"id": "c5598ebc-5665-48b1-8d08-dcd4b12eb9f4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-945495745", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f75cf3b-ac", "ovs_interfaceid": "7f75cf3b-acb2-456d-896e-25184e6811c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d1e8b347-53e9-4aeb-82d6-f9e255d24442", "address": "fa:16:3e:28:04:2c", "network": {"id": "cd0bbfa3-9d50-4c94-b6ef-8d2e4506d27d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-203381185", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.167", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1e8b347-53", "ovs_interfaceid": "d1e8b347-53e9-4aeb-82d6-f9e255d24442", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a52a68b2-6f6c-4b4d-b377-b79e60f46dac", "address": "fa:16:3e:43:07:3d", "network": {"id": "c5598ebc-5665-48b1-8d08-dcd4b12eb9f4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-945495745", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.129", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa52a68b2-6f", "ovs_interfaceid": "a52a68b2-6f6c-4b4d-b377-b79e60f46dac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.448248] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': task-1023697, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066094} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.448562] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 658.449441] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049325a2-0a79-4c49-8787-91e6221a42fc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.473353] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Reconfiguring VM instance instance-0000001c to attach disk [datastore2] 779e3559-37a5-49b2-b17a-bf95d0d8d23c/779e3559-37a5-49b2-b17a-bf95d0d8d23c.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 658.475293] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2bc4319d-ed60-40dd-89c8-c4ec1dcab027 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.492141] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.492642] env[63372]: DEBUG nova.compute.manager [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 658.495638] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.255s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.497054] env[63372]: INFO nova.compute.claims [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 658.505126] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Waiting for the task: (returnval){ [ 658.505126] env[63372]: value = "task-1023701" [ 658.505126] env[63372]: _type = "Task" [ 658.505126] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.514502] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': task-1023701, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.526991] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023700, 'name': CreateVM_Task, 'duration_secs': 0.324222} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.527200] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 658.528256] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.528506] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.528832] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 658.531246] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abedbcb4-c271-4dd5-8d01-3cfb9daa7993 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.536054] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Waiting for the task: (returnval){ [ 658.536054] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529eb5e1-11f3-ef23-6dc4-de8b39ecce58" [ 658.536054] env[63372]: _type = "Task" [ 658.536054] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.544576] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529eb5e1-11f3-ef23-6dc4-de8b39ecce58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.814730] env[63372]: DEBUG nova.network.neutron [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Updated VIF entry in instance network info cache for port e56add8e-4342-4ab4-b2f2-c03f257f3d3a. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 658.815100] env[63372]: DEBUG nova.network.neutron [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Updating instance_info_cache with network_info: [{"id": "e56add8e-4342-4ab4-b2f2-c03f257f3d3a", "address": "fa:16:3e:11:78:8c", "network": {"id": "79050166-5dc7-4ef3-9819-ac534e041985", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-2103555585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06cecd1557de407697ad403a85138481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24727047-6358-4015-86c1-394ab07fb88f", "external-id": "nsx-vlan-transportzone-476", "segmentation_id": 476, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape56add8e-43", "ovs_interfaceid": "e56add8e-4342-4ab4-b2f2-c03f257f3d3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.894346] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.894346] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.894346] env[63372]: DEBUG nova.network.neutron [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 658.932513] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Releasing lock "refresh_cache-6d4502b7-b9df-4fef-8400-9ff07b820e2b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.933837] env[63372]: DEBUG nova.compute.manager [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Instance network_info: |[{"id": "7f75cf3b-acb2-456d-896e-25184e6811c4", "address": "fa:16:3e:f0:dc:5b", "network": {"id": "c5598ebc-5665-48b1-8d08-dcd4b12eb9f4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-945495745", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f75cf3b-ac", "ovs_interfaceid": "7f75cf3b-acb2-456d-896e-25184e6811c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d1e8b347-53e9-4aeb-82d6-f9e255d24442", "address": "fa:16:3e:28:04:2c", "network": {"id": "cd0bbfa3-9d50-4c94-b6ef-8d2e4506d27d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-203381185", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.167", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1e8b347-53", "ovs_interfaceid": "d1e8b347-53e9-4aeb-82d6-f9e255d24442", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a52a68b2-6f6c-4b4d-b377-b79e60f46dac", "address": "fa:16:3e:43:07:3d", "network": {"id": "c5598ebc-5665-48b1-8d08-dcd4b12eb9f4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-945495745", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.129", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa52a68b2-6f", "ovs_interfaceid": "a52a68b2-6f6c-4b4d-b377-b79e60f46dac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 658.933837] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] Acquired lock "refresh_cache-6d4502b7-b9df-4fef-8400-9ff07b820e2b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.933837] env[63372]: DEBUG nova.network.neutron [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Refreshing network info cache for port d1e8b347-53e9-4aeb-82d6-f9e255d24442 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 658.934323] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:dc:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7f75cf3b-acb2-456d-896e-25184e6811c4', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:04:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b107fab-ee71-47db-ad4d-3c6f05546843', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd1e8b347-53e9-4aeb-82d6-f9e255d24442', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:07:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a52a68b2-6f6c-4b4d-b377-b79e60f46dac', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 658.945619] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Creating folder: Project (056172ef521c4de08b5033057773b2ac). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 658.948651] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4dc5ab49-1153-496a-99f1-0fbc847e4cf8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.960170] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Created folder: Project (056172ef521c4de08b5033057773b2ac) in parent group-v227230. [ 658.960400] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Creating folder: Instances. Parent ref: group-v227249. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 658.960675] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9d624eb5-e1c0-44d8-a2e5-b83cd060fff7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.969100] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Created folder: Instances in parent group-v227249. [ 658.969320] env[63372]: DEBUG oslo.service.loopingcall [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.969496] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 658.970101] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-60115cec-8f32-4330-9638-85e4e2458ab3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.992554] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 658.992554] env[63372]: value = "task-1023704" [ 658.992554] env[63372]: _type = "Task" [ 658.992554] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.001063] env[63372]: DEBUG nova.compute.utils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 659.003946] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023704, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.006580] env[63372]: DEBUG nova.compute.manager [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 659.006775] env[63372]: DEBUG nova.network.neutron [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 659.018640] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': task-1023701, 'name': ReconfigVM_Task, 'duration_secs': 0.299918} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.018640] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Reconfigured VM instance instance-0000001c to attach disk [datastore2] 779e3559-37a5-49b2-b17a-bf95d0d8d23c/779e3559-37a5-49b2-b17a-bf95d0d8d23c.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 659.019422] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de182f93-8d29-4b51-9b9f-ace5b5eb8671 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.025720] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Waiting for the task: (returnval){ [ 659.025720] env[63372]: value = "task-1023705" [ 659.025720] env[63372]: _type = "Task" [ 659.025720] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.034724] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': task-1023705, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.045118] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529eb5e1-11f3-ef23-6dc4-de8b39ecce58, 'name': SearchDatastore_Task, 'duration_secs': 0.008432} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.045388] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.045638] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 659.045937] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.046116] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.046301] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 659.046545] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4d2c4db4-998c-48e8-a924-20cba106d37e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.052896] env[63372]: DEBUG nova.policy [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f88a1b3e87c94bccb589e8ef24f6efd6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'baadac70098a43fab6dfb45f6a043548', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 659.055933] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 659.056151] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 659.057213] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29f8c15c-ba76-4e1c-91b7-ade198e8984c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.062481] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Waiting for the task: (returnval){ [ 659.062481] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52969f90-3663-f80d-24a0-99107d34f9a7" [ 659.062481] env[63372]: _type = "Task" [ 659.062481] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.070570] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52969f90-3663-f80d-24a0-99107d34f9a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.318571] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] Releasing lock "refresh_cache-cf673ac1-2c7d-468b-83ec-c723d5182457" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.318853] env[63372]: DEBUG nova.compute.manager [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Received event network-changed-a52a68b2-6f6c-4b4d-b377-b79e60f46dac {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 659.319032] env[63372]: DEBUG nova.compute.manager [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Refreshing instance network info cache due to event network-changed-a52a68b2-6f6c-4b4d-b377-b79e60f46dac. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 659.319224] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] Acquiring lock "refresh_cache-6d4502b7-b9df-4fef-8400-9ff07b820e2b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.322999] env[63372]: DEBUG nova.network.neutron [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Updated VIF entry in instance network info cache for port d1e8b347-53e9-4aeb-82d6-f9e255d24442. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 659.323226] env[63372]: DEBUG nova.network.neutron [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Updating instance_info_cache with network_info: [{"id": "7f75cf3b-acb2-456d-896e-25184e6811c4", "address": "fa:16:3e:f0:dc:5b", "network": {"id": "c5598ebc-5665-48b1-8d08-dcd4b12eb9f4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-945495745", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f75cf3b-ac", "ovs_interfaceid": "7f75cf3b-acb2-456d-896e-25184e6811c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d1e8b347-53e9-4aeb-82d6-f9e255d24442", "address": "fa:16:3e:28:04:2c", "network": {"id": "cd0bbfa3-9d50-4c94-b6ef-8d2e4506d27d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-203381185", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.167", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1e8b347-53", "ovs_interfaceid": "d1e8b347-53e9-4aeb-82d6-f9e255d24442", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a52a68b2-6f6c-4b4d-b377-b79e60f46dac", "address": "fa:16:3e:43:07:3d", "network": {"id": "c5598ebc-5665-48b1-8d08-dcd4b12eb9f4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-945495745", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.129", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa52a68b2-6f", "ovs_interfaceid": "a52a68b2-6f6c-4b4d-b377-b79e60f46dac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.335138] env[63372]: DEBUG nova.network.neutron [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Successfully created port: fd39de6f-7b40-4e7b-a88c-7996e802c774 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 659.425546] env[63372]: DEBUG nova.network.neutron [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.509030] env[63372]: DEBUG nova.compute.manager [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 659.510792] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023704, 'name': CreateVM_Task, 'duration_secs': 0.48751} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.513290] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 659.514497] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.514659] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.515028] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 659.515595] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11f971b8-d40e-4bb3-a06a-4746087210e0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.519901] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 659.519901] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f0f675-d66c-c985-7c7a-bd3128fe0ffe" [ 659.519901] env[63372]: _type = "Task" [ 659.519901] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.528169] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f0f675-d66c-c985-7c7a-bd3128fe0ffe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.535957] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': task-1023705, 'name': Rename_Task, 'duration_secs': 0.148314} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.536713] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 659.536713] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e7015d70-b03d-458b-8452-03f211e9d4e3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.544099] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Waiting for the task: (returnval){ [ 659.544099] env[63372]: value = "task-1023706" [ 659.544099] env[63372]: _type = "Task" [ 659.544099] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.554274] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': task-1023706, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.572328] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52969f90-3663-f80d-24a0-99107d34f9a7, 'name': SearchDatastore_Task, 'duration_secs': 0.017712} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.575849] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb3e0e19-bc0e-40f7-ade9-d8bcf8cf3441 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.581455] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Waiting for the task: (returnval){ [ 659.581455] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529f495c-2395-dec4-1024-d2b7c592b770" [ 659.581455] env[63372]: _type = "Task" [ 659.581455] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.591916] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529f495c-2395-dec4-1024-d2b7c592b770, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.656766] env[63372]: DEBUG nova.network.neutron [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Updating instance_info_cache with network_info: [{"id": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "address": "fa:16:3e:17:c8:2e", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4cf27a3-fc", "ovs_interfaceid": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.827440] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d156649-dd9d-435a-90de-93adf97b8821 req-4ae5a6c6-92d8-4e66-8b6c-0f5b9b50773e service nova] Releasing lock "refresh_cache-6d4502b7-b9df-4fef-8400-9ff07b820e2b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.827918] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] Acquired lock "refresh_cache-6d4502b7-b9df-4fef-8400-9ff07b820e2b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.828156] env[63372]: DEBUG nova.network.neutron [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Refreshing network info cache for port a52a68b2-6f6c-4b4d-b377-b79e60f46dac {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 659.987789] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f33ea72-1aac-4f24-8f1d-b9159615b04b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.996326] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52463349-c5c0-424b-8021-4a40bd7bf61d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.035844] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210a55d0-d8a8-416d-a474-039bb0f2acb7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.039204] env[63372]: DEBUG nova.compute.manager [req-b61c1618-37e3-4f8c-b0c0-5bc74f3d3a2f req-079ac4c4-8749-4a97-9a3d-9e4d2e5f3a52 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Received event network-vif-plugged-b4cf27a3-fc17-4ac5-b96a-3be058f71e03 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 660.039403] env[63372]: DEBUG oslo_concurrency.lockutils [req-b61c1618-37e3-4f8c-b0c0-5bc74f3d3a2f req-079ac4c4-8749-4a97-9a3d-9e4d2e5f3a52 service nova] Acquiring lock "bbba9f28-045d-41ab-8539-5b2968fe3d54-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.039603] env[63372]: DEBUG oslo_concurrency.lockutils [req-b61c1618-37e3-4f8c-b0c0-5bc74f3d3a2f req-079ac4c4-8749-4a97-9a3d-9e4d2e5f3a52 service nova] Lock "bbba9f28-045d-41ab-8539-5b2968fe3d54-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.039768] env[63372]: DEBUG oslo_concurrency.lockutils [req-b61c1618-37e3-4f8c-b0c0-5bc74f3d3a2f req-079ac4c4-8749-4a97-9a3d-9e4d2e5f3a52 service nova] Lock "bbba9f28-045d-41ab-8539-5b2968fe3d54-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.039926] env[63372]: DEBUG nova.compute.manager [req-b61c1618-37e3-4f8c-b0c0-5bc74f3d3a2f req-079ac4c4-8749-4a97-9a3d-9e4d2e5f3a52 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] No waiting events found dispatching network-vif-plugged-b4cf27a3-fc17-4ac5-b96a-3be058f71e03 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 660.040118] env[63372]: WARNING nova.compute.manager [req-b61c1618-37e3-4f8c-b0c0-5bc74f3d3a2f req-079ac4c4-8749-4a97-9a3d-9e4d2e5f3a52 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Received unexpected event network-vif-plugged-b4cf27a3-fc17-4ac5-b96a-3be058f71e03 for instance with vm_state building and task_state spawning. [ 660.040288] env[63372]: DEBUG nova.compute.manager [req-b61c1618-37e3-4f8c-b0c0-5bc74f3d3a2f req-079ac4c4-8749-4a97-9a3d-9e4d2e5f3a52 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Received event network-changed-b4cf27a3-fc17-4ac5-b96a-3be058f71e03 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 660.040435] env[63372]: DEBUG nova.compute.manager [req-b61c1618-37e3-4f8c-b0c0-5bc74f3d3a2f req-079ac4c4-8749-4a97-9a3d-9e4d2e5f3a52 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Refreshing instance network info cache due to event network-changed-b4cf27a3-fc17-4ac5-b96a-3be058f71e03. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 660.040596] env[63372]: DEBUG oslo_concurrency.lockutils [req-b61c1618-37e3-4f8c-b0c0-5bc74f3d3a2f req-079ac4c4-8749-4a97-9a3d-9e4d2e5f3a52 service nova] Acquiring lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.052159] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733a1346-ccf4-4325-9a79-5fa5aec8eab4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.055901] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f0f675-d66c-c985-7c7a-bd3128fe0ffe, 'name': SearchDatastore_Task, 'duration_secs': 0.008406} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.056440] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.056662] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 660.056863] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.061951] env[63372]: DEBUG oslo_vmware.api [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': task-1023706, 'name': PowerOnVM_Task, 'duration_secs': 0.468092} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.069952] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 660.070189] env[63372]: INFO nova.compute.manager [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Took 7.05 seconds to spawn the instance on the hypervisor. [ 660.070363] env[63372]: DEBUG nova.compute.manager [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 660.071242] env[63372]: DEBUG nova.compute.provider_tree [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.072736] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62dd934-ad73-4684-bfdf-9c309954ae07 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.091179] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529f495c-2395-dec4-1024-d2b7c592b770, 'name': SearchDatastore_Task, 'duration_secs': 0.009256} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.091423] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.091664] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] cf673ac1-2c7d-468b-83ec-c723d5182457/cf673ac1-2c7d-468b-83ec-c723d5182457.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 660.091929] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.092125] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 660.092323] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bfeb1305-d2c4-4a46-a09c-49bd0b998b68 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.094306] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fd72b907-9615-4fd0-9728-7b32388a1adf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.100389] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Waiting for the task: (returnval){ [ 660.100389] env[63372]: value = "task-1023707" [ 660.100389] env[63372]: _type = "Task" [ 660.100389] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.104643] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 660.104820] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 660.106019] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e3ee376-b7d6-4480-a9b6-0414727533aa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.111466] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1023707, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.114433] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 660.114433] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52aa7cd0-df1e-2cd8-47c9-f2a89e6b84c4" [ 660.114433] env[63372]: _type = "Task" [ 660.114433] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.122753] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52aa7cd0-df1e-2cd8-47c9-f2a89e6b84c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.161089] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.161499] env[63372]: DEBUG nova.compute.manager [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Instance network_info: |[{"id": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "address": "fa:16:3e:17:c8:2e", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4cf27a3-fc", "ovs_interfaceid": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 660.161820] env[63372]: DEBUG oslo_concurrency.lockutils [req-b61c1618-37e3-4f8c-b0c0-5bc74f3d3a2f req-079ac4c4-8749-4a97-9a3d-9e4d2e5f3a52 service nova] Acquired lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.162039] env[63372]: DEBUG nova.network.neutron [req-b61c1618-37e3-4f8c-b0c0-5bc74f3d3a2f req-079ac4c4-8749-4a97-9a3d-9e4d2e5f3a52 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Refreshing network info cache for port b4cf27a3-fc17-4ac5-b96a-3be058f71e03 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 660.163376] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:c8:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b4cf27a3-fc17-4ac5-b96a-3be058f71e03', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 660.171527] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Creating folder: Project (46d2cb188a004dc385f2aa15458be929). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 660.174641] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3191ea2b-528d-4277-8b01-756623cb71b5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.186021] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Created folder: Project (46d2cb188a004dc385f2aa15458be929) in parent group-v227230. [ 660.186021] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Creating folder: Instances. Parent ref: group-v227252. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 660.186021] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e54794c5-0e30-4643-9423-4a8804f25739 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.195952] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Created folder: Instances in parent group-v227252. [ 660.196202] env[63372]: DEBUG oslo.service.loopingcall [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 660.196440] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 660.196580] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f2f13c21-5d1a-446c-b95d-134a0e5ca927 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.217609] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 660.217609] env[63372]: value = "task-1023710" [ 660.217609] env[63372]: _type = "Task" [ 660.217609] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.225389] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023710, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.507858] env[63372]: DEBUG nova.network.neutron [req-b61c1618-37e3-4f8c-b0c0-5bc74f3d3a2f req-079ac4c4-8749-4a97-9a3d-9e4d2e5f3a52 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Updated VIF entry in instance network info cache for port b4cf27a3-fc17-4ac5-b96a-3be058f71e03. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 660.507858] env[63372]: DEBUG nova.network.neutron [req-b61c1618-37e3-4f8c-b0c0-5bc74f3d3a2f req-079ac4c4-8749-4a97-9a3d-9e4d2e5f3a52 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Updating instance_info_cache with network_info: [{"id": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "address": "fa:16:3e:17:c8:2e", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4cf27a3-fc", "ovs_interfaceid": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.533878] env[63372]: DEBUG nova.compute.manager [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 660.562120] env[63372]: DEBUG nova.virt.hardware [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 660.562377] env[63372]: DEBUG nova.virt.hardware [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 660.562530] env[63372]: DEBUG nova.virt.hardware [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 660.562852] env[63372]: DEBUG nova.virt.hardware [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 660.562852] env[63372]: DEBUG nova.virt.hardware [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 660.562993] env[63372]: DEBUG nova.virt.hardware [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 660.563228] env[63372]: DEBUG nova.virt.hardware [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 660.563344] env[63372]: DEBUG nova.virt.hardware [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 660.563506] env[63372]: DEBUG nova.virt.hardware [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 660.563689] env[63372]: DEBUG nova.virt.hardware [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 660.563871] env[63372]: DEBUG nova.virt.hardware [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 660.564768] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac0d836-7764-449a-b7d7-8575403cac1c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.572984] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284639a1-e9ed-4149-b480-35f736ec688a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.578230] env[63372]: DEBUG nova.scheduler.client.report [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 660.603346] env[63372]: DEBUG nova.network.neutron [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Updated VIF entry in instance network info cache for port a52a68b2-6f6c-4b4d-b377-b79e60f46dac. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 660.603346] env[63372]: DEBUG nova.network.neutron [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Updating instance_info_cache with network_info: [{"id": "7f75cf3b-acb2-456d-896e-25184e6811c4", "address": "fa:16:3e:f0:dc:5b", "network": {"id": "c5598ebc-5665-48b1-8d08-dcd4b12eb9f4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-945495745", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f75cf3b-ac", "ovs_interfaceid": "7f75cf3b-acb2-456d-896e-25184e6811c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d1e8b347-53e9-4aeb-82d6-f9e255d24442", "address": "fa:16:3e:28:04:2c", "network": {"id": "cd0bbfa3-9d50-4c94-b6ef-8d2e4506d27d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-203381185", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.167", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1e8b347-53", "ovs_interfaceid": "d1e8b347-53e9-4aeb-82d6-f9e255d24442", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a52a68b2-6f6c-4b4d-b377-b79e60f46dac", "address": "fa:16:3e:43:07:3d", "network": {"id": "c5598ebc-5665-48b1-8d08-dcd4b12eb9f4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-945495745", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.129", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa52a68b2-6f", "ovs_interfaceid": "a52a68b2-6f6c-4b4d-b377-b79e60f46dac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.605103] env[63372]: INFO nova.compute.manager [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Took 38.85 seconds to build instance. [ 660.616595] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1023707, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.627483] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52aa7cd0-df1e-2cd8-47c9-f2a89e6b84c4, 'name': SearchDatastore_Task, 'duration_secs': 0.028593} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.628530] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b758ddb-196a-4544-869b-0490e10b8d2a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.633856] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 660.633856] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523605e8-ae48-39e5-11ec-3bbd30b7b0a1" [ 660.633856] env[63372]: _type = "Task" [ 660.633856] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.642283] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523605e8-ae48-39e5-11ec-3bbd30b7b0a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.727760] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023710, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.842071] env[63372]: DEBUG nova.network.neutron [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Successfully updated port: fd39de6f-7b40-4e7b-a88c-7996e802c774 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 661.010658] env[63372]: DEBUG oslo_concurrency.lockutils [req-b61c1618-37e3-4f8c-b0c0-5bc74f3d3a2f req-079ac4c4-8749-4a97-9a3d-9e4d2e5f3a52 service nova] Releasing lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.086022] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.588s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.086022] env[63372]: DEBUG nova.compute.manager [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 661.087108] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.351s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.106696] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecb71d1c-5760-4b81-9b53-bc9035d3edf9 req-7a8cfe9b-cc17-4b7f-b9d2-f0f1386635ec service nova] Releasing lock "refresh_cache-6d4502b7-b9df-4fef-8400-9ff07b820e2b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.110975] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c902dee7-7664-407e-b310-40abdfe43f81 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Lock "779e3559-37a5-49b2-b17a-bf95d0d8d23c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.521s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.116790] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1023707, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545384} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.117267] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] cf673ac1-2c7d-468b-83ec-c723d5182457/cf673ac1-2c7d-468b-83ec-c723d5182457.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 661.117472] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 661.117708] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7a67c9b6-726e-4074-87cd-9e618d102f5a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.126506] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Waiting for the task: (returnval){ [ 661.126506] env[63372]: value = "task-1023711" [ 661.126506] env[63372]: _type = "Task" [ 661.126506] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.138196] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1023711, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.150784] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523605e8-ae48-39e5-11ec-3bbd30b7b0a1, 'name': SearchDatastore_Task, 'duration_secs': 0.008993} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.151072] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.151355] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 6d4502b7-b9df-4fef-8400-9ff07b820e2b/6d4502b7-b9df-4fef-8400-9ff07b820e2b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 661.151616] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1dc00088-3a61-438b-a0e0-613c2ebe7c6a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.158203] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 661.158203] env[63372]: value = "task-1023712" [ 661.158203] env[63372]: _type = "Task" [ 661.158203] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.167536] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023712, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.228195] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023710, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.345646] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Acquiring lock "refresh_cache-9404a0f0-a3e5-43ee-bc20-e9566cfe5a60" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.345834] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Acquired lock "refresh_cache-9404a0f0-a3e5-43ee-bc20-e9566cfe5a60" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.345994] env[63372]: DEBUG nova.network.neutron [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 661.422798] env[63372]: DEBUG oslo_concurrency.lockutils [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Acquiring lock "779e3559-37a5-49b2-b17a-bf95d0d8d23c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.423063] env[63372]: DEBUG oslo_concurrency.lockutils [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Lock "779e3559-37a5-49b2-b17a-bf95d0d8d23c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.423274] env[63372]: DEBUG oslo_concurrency.lockutils [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Acquiring lock "779e3559-37a5-49b2-b17a-bf95d0d8d23c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.423449] env[63372]: DEBUG oslo_concurrency.lockutils [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Lock "779e3559-37a5-49b2-b17a-bf95d0d8d23c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.423765] env[63372]: DEBUG oslo_concurrency.lockutils [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Lock "779e3559-37a5-49b2-b17a-bf95d0d8d23c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.425897] env[63372]: INFO nova.compute.manager [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Terminating instance [ 661.427683] env[63372]: DEBUG nova.compute.manager [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 661.427882] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 661.428774] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03564a86-ff34-4d75-81ec-ae9152adf0cf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.436594] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 661.436836] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e80dad55-7c9a-43aa-ad3b-b74f6e1ce2f3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.443457] env[63372]: DEBUG oslo_vmware.api [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Waiting for the task: (returnval){ [ 661.443457] env[63372]: value = "task-1023713" [ 661.443457] env[63372]: _type = "Task" [ 661.443457] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.451278] env[63372]: DEBUG oslo_vmware.api [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': task-1023713, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.592457] env[63372]: DEBUG nova.compute.utils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 661.593922] env[63372]: DEBUG nova.compute.manager [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 661.594121] env[63372]: DEBUG nova.network.neutron [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 661.618638] env[63372]: DEBUG nova.compute.manager [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 661.637490] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1023711, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06977} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.637935] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 661.639040] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6aeed12-add9-4aad-bf2e-1634251844c1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.670518] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Reconfiguring VM instance instance-0000001d to attach disk [datastore2] cf673ac1-2c7d-468b-83ec-c723d5182457/cf673ac1-2c7d-468b-83ec-c723d5182457.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 661.674862] env[63372]: DEBUG nova.policy [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab5ce48df43d44ac9812cf3925c1fab6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c8b350f69f6d43ec869511e0e8fdb98c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 661.681028] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18bea1c4-84a2-4410-bd9f-8f03b030159b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.702586] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023712, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.704367] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Waiting for the task: (returnval){ [ 661.704367] env[63372]: value = "task-1023714" [ 661.704367] env[63372]: _type = "Task" [ 661.704367] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.715655] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1023714, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.729546] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023710, 'name': CreateVM_Task, 'duration_secs': 1.490187} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.729722] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 661.730427] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.730591] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.730916] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 661.731209] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b53fea52-8d6e-4204-85bb-7e416cdc5c95 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.738775] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 661.738775] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52ac1aa4-6019-81f6-0809-39270a3dc847" [ 661.738775] env[63372]: _type = "Task" [ 661.738775] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.746783] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ac1aa4-6019-81f6-0809-39270a3dc847, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.902911] env[63372]: DEBUG nova.network.neutron [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.960568] env[63372]: DEBUG oslo_vmware.api [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': task-1023713, 'name': PowerOffVM_Task, 'duration_secs': 0.274695} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.960568] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 661.960568] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 661.960788] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-adbd1625-2a04-4c94-a348-24a0a508e015 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.037133] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 662.037404] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 662.037589] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Deleting the datastore file [datastore2] 779e3559-37a5-49b2-b17a-bf95d0d8d23c {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 662.037865] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-08ebc393-5cef-4c4f-88d5-00d7027f8c29 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.046823] env[63372]: DEBUG oslo_vmware.api [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Waiting for the task: (returnval){ [ 662.046823] env[63372]: value = "task-1023716" [ 662.046823] env[63372]: _type = "Task" [ 662.046823] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.059619] env[63372]: DEBUG oslo_vmware.api [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': task-1023716, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.081959] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75d2bf02-2741-45a6-9b87-a96710cceba5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.087211] env[63372]: DEBUG nova.compute.manager [req-d543cdea-e68d-409e-861d-d5b75e75f81e req-6ad79ada-619f-459d-acec-bb7223e29cdf service nova] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Received event network-vif-plugged-fd39de6f-7b40-4e7b-a88c-7996e802c774 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 662.087421] env[63372]: DEBUG oslo_concurrency.lockutils [req-d543cdea-e68d-409e-861d-d5b75e75f81e req-6ad79ada-619f-459d-acec-bb7223e29cdf service nova] Acquiring lock "9404a0f0-a3e5-43ee-bc20-e9566cfe5a60-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.087612] env[63372]: DEBUG oslo_concurrency.lockutils [req-d543cdea-e68d-409e-861d-d5b75e75f81e req-6ad79ada-619f-459d-acec-bb7223e29cdf service nova] Lock "9404a0f0-a3e5-43ee-bc20-e9566cfe5a60-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.087773] env[63372]: DEBUG oslo_concurrency.lockutils [req-d543cdea-e68d-409e-861d-d5b75e75f81e req-6ad79ada-619f-459d-acec-bb7223e29cdf service nova] Lock "9404a0f0-a3e5-43ee-bc20-e9566cfe5a60-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.087937] env[63372]: DEBUG nova.compute.manager [req-d543cdea-e68d-409e-861d-d5b75e75f81e req-6ad79ada-619f-459d-acec-bb7223e29cdf service nova] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] No waiting events found dispatching network-vif-plugged-fd39de6f-7b40-4e7b-a88c-7996e802c774 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 662.088108] env[63372]: WARNING nova.compute.manager [req-d543cdea-e68d-409e-861d-d5b75e75f81e req-6ad79ada-619f-459d-acec-bb7223e29cdf service nova] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Received unexpected event network-vif-plugged-fd39de6f-7b40-4e7b-a88c-7996e802c774 for instance with vm_state building and task_state spawning. [ 662.088266] env[63372]: DEBUG nova.compute.manager [req-d543cdea-e68d-409e-861d-d5b75e75f81e req-6ad79ada-619f-459d-acec-bb7223e29cdf service nova] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Received event network-changed-fd39de6f-7b40-4e7b-a88c-7996e802c774 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 662.088413] env[63372]: DEBUG nova.compute.manager [req-d543cdea-e68d-409e-861d-d5b75e75f81e req-6ad79ada-619f-459d-acec-bb7223e29cdf service nova] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Refreshing instance network info cache due to event network-changed-fd39de6f-7b40-4e7b-a88c-7996e802c774. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 662.088574] env[63372]: DEBUG oslo_concurrency.lockutils [req-d543cdea-e68d-409e-861d-d5b75e75f81e req-6ad79ada-619f-459d-acec-bb7223e29cdf service nova] Acquiring lock "refresh_cache-9404a0f0-a3e5-43ee-bc20-e9566cfe5a60" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.094070] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72b695a-0ab1-4a13-a6aa-61d5b50ae7a6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.103385] env[63372]: DEBUG nova.compute.manager [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 662.152547] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea42c0f3-b7e6-48df-ac32-2d003e7c06eb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.152547] env[63372]: DEBUG nova.network.neutron [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Successfully created port: 7b5b0928-bb42-4396-bf97-ad9cb3423a7a {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 662.161480] env[63372]: DEBUG nova.network.neutron [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Updating instance_info_cache with network_info: [{"id": "fd39de6f-7b40-4e7b-a88c-7996e802c774", "address": "fa:16:3e:21:de:13", "network": {"id": "07f988a8-80e0-4a49-a8c1-df19313a423c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-2017897237-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "baadac70098a43fab6dfb45f6a043548", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f85835c8-5d0c-4b2f-97c4-6c4006580f79", "external-id": "nsx-vlan-transportzone-245", "segmentation_id": 245, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd39de6f-7b", "ovs_interfaceid": "fd39de6f-7b40-4e7b-a88c-7996e802c774", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.164716] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-806b7a41-f7c2-4a96-b84d-d7f70e9b1812 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.170191] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.184665] env[63372]: DEBUG nova.compute.provider_tree [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.189918] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023712, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.969227} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.190341] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 6d4502b7-b9df-4fef-8400-9ff07b820e2b/6d4502b7-b9df-4fef-8400-9ff07b820e2b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 662.190782] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 662.190782] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-218031b8-3631-4f54-bd83-67889012c972 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.201028] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 662.201028] env[63372]: value = "task-1023717" [ 662.201028] env[63372]: _type = "Task" [ 662.201028] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.213668] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023717, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.220166] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1023714, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.249114] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ac1aa4-6019-81f6-0809-39270a3dc847, 'name': SearchDatastore_Task, 'duration_secs': 0.009789} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.249417] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.249644] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 662.249871] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.250017] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.250194] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 662.250446] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cf7f8f3c-7b9f-42d6-a2ad-dadbdb06d7a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.258700] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 662.258925] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 662.259753] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c6c6d91-b647-4670-8a6e-26999442df09 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.265446] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 662.265446] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]526d4750-7343-c795-baee-9772558e652b" [ 662.265446] env[63372]: _type = "Task" [ 662.265446] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.273212] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526d4750-7343-c795-baee-9772558e652b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.556548] env[63372]: DEBUG oslo_vmware.api [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Task: {'id': task-1023716, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.22994} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.556840] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 662.557085] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 662.557217] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 662.557386] env[63372]: INFO nova.compute.manager [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Took 1.13 seconds to destroy the instance on the hypervisor. [ 662.557621] env[63372]: DEBUG oslo.service.loopingcall [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 662.557808] env[63372]: DEBUG nova.compute.manager [-] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.557904] env[63372]: DEBUG nova.network.neutron [-] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 662.670532] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Releasing lock "refresh_cache-9404a0f0-a3e5-43ee-bc20-e9566cfe5a60" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.670880] env[63372]: DEBUG nova.compute.manager [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Instance network_info: |[{"id": "fd39de6f-7b40-4e7b-a88c-7996e802c774", "address": "fa:16:3e:21:de:13", "network": {"id": "07f988a8-80e0-4a49-a8c1-df19313a423c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-2017897237-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "baadac70098a43fab6dfb45f6a043548", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f85835c8-5d0c-4b2f-97c4-6c4006580f79", "external-id": "nsx-vlan-transportzone-245", "segmentation_id": 245, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd39de6f-7b", "ovs_interfaceid": "fd39de6f-7b40-4e7b-a88c-7996e802c774", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 662.671209] env[63372]: DEBUG oslo_concurrency.lockutils [req-d543cdea-e68d-409e-861d-d5b75e75f81e req-6ad79ada-619f-459d-acec-bb7223e29cdf service nova] Acquired lock "refresh_cache-9404a0f0-a3e5-43ee-bc20-e9566cfe5a60" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.671387] env[63372]: DEBUG nova.network.neutron [req-d543cdea-e68d-409e-861d-d5b75e75f81e req-6ad79ada-619f-459d-acec-bb7223e29cdf service nova] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Refreshing network info cache for port fd39de6f-7b40-4e7b-a88c-7996e802c774 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 662.672535] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:de:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f85835c8-5d0c-4b2f-97c4-6c4006580f79', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fd39de6f-7b40-4e7b-a88c-7996e802c774', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 662.680751] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Creating folder: Project (baadac70098a43fab6dfb45f6a043548). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 662.681446] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-66898794-f596-4a0b-99c4-ad26540f2fcb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.691773] env[63372]: DEBUG nova.scheduler.client.report [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.696229] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Created folder: Project (baadac70098a43fab6dfb45f6a043548) in parent group-v227230. [ 662.696413] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Creating folder: Instances. Parent ref: group-v227255. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 662.696815] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a130d644-b21e-46ac-9aba-542f6c82919a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.708229] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023717, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065613} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.712364] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 662.712639] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Created folder: Instances in parent group-v227255. [ 662.712848] env[63372]: DEBUG oslo.service.loopingcall [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 662.713704] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771c8abf-49ee-4c9e-a5eb-1128200cab5a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.715887] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 662.716407] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9fefaa61-5b4e-4aae-83cf-55a5f2bfe5a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.733409] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1023714, 'name': ReconfigVM_Task, 'duration_secs': 0.621865} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.733967] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Reconfigured VM instance instance-0000001d to attach disk [datastore2] cf673ac1-2c7d-468b-83ec-c723d5182457/cf673ac1-2c7d-468b-83ec-c723d5182457.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 662.735496] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c9bd68f-4ea8-4ca4-8385-d947b01c1bc1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.757852] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Reconfiguring VM instance instance-0000001b to attach disk [datastore2] 6d4502b7-b9df-4fef-8400-9ff07b820e2b/6d4502b7-b9df-4fef-8400-9ff07b820e2b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 662.759709] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad34ac36-b86a-4923-90bb-3bd28a9cee5e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.774137] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 662.774137] env[63372]: value = "task-1023720" [ 662.774137] env[63372]: _type = "Task" [ 662.774137] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.777257] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Waiting for the task: (returnval){ [ 662.777257] env[63372]: value = "task-1023721" [ 662.777257] env[63372]: _type = "Task" [ 662.777257] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.789274] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 662.789274] env[63372]: value = "task-1023722" [ 662.789274] env[63372]: _type = "Task" [ 662.789274] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.801627] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023720, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.801928] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526d4750-7343-c795-baee-9772558e652b, 'name': SearchDatastore_Task, 'duration_secs': 0.01636} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.804543] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39612f80-d131-4191-8994-7def3fe4aaa8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.810643] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023722, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.814117] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1023721, 'name': Rename_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.816368] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 662.816368] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52a78d8f-4e86-c8fd-6ac6-2586ea487a4e" [ 662.816368] env[63372]: _type = "Task" [ 662.816368] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.824131] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a78d8f-4e86-c8fd-6ac6-2586ea487a4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.111451] env[63372]: DEBUG nova.compute.manager [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 663.139017] env[63372]: DEBUG nova.virt.hardware [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 663.139212] env[63372]: DEBUG nova.virt.hardware [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 663.139420] env[63372]: DEBUG nova.virt.hardware [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 663.139679] env[63372]: DEBUG nova.virt.hardware [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 663.139890] env[63372]: DEBUG nova.virt.hardware [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 663.140111] env[63372]: DEBUG nova.virt.hardware [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 663.140354] env[63372]: DEBUG nova.virt.hardware [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 663.140601] env[63372]: DEBUG nova.virt.hardware [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 663.140870] env[63372]: DEBUG nova.virt.hardware [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 663.141119] env[63372]: DEBUG nova.virt.hardware [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 663.141359] env[63372]: DEBUG nova.virt.hardware [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 663.142411] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-930cfcca-81e4-49c0-970c-aa2948916f7c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.152170] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af074060-111d-4ac1-88a6-2610a6dcdf7f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.199061] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.112s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.199741] env[63372]: ERROR nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 61938f95-e588-4bdf-b61a-0795bd8279e3, please check neutron logs for more information. [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Traceback (most recent call last): [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] self.driver.spawn(context, instance, image_meta, [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] vm_ref = self.build_virtual_machine(instance, [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] for vif in network_info: [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] return self._sync_wrapper(fn, *args, **kwargs) [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] self.wait() [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] self[:] = self._gt.wait() [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] return self._exit_event.wait() [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] result = hub.switch() [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] return self.greenlet.switch() [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] result = function(*args, **kwargs) [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] return func(*args, **kwargs) [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] raise e [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] nwinfo = self.network_api.allocate_for_instance( [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] created_port_ids = self._update_ports_for_instance( [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] with excutils.save_and_reraise_exception(): [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.199741] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] self.force_reraise() [ 663.202106] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.202106] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] raise self.value [ 663.202106] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.202106] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] updated_port = self._update_port( [ 663.202106] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.202106] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] _ensure_no_port_binding_failure(port) [ 663.202106] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.202106] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] raise exception.PortBindingFailed(port_id=port['id']) [ 663.202106] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] nova.exception.PortBindingFailed: Binding failed for port 61938f95-e588-4bdf-b61a-0795bd8279e3, please check neutron logs for more information. [ 663.202106] env[63372]: ERROR nova.compute.manager [instance: ae86a977-0a9a-46e5-9072-cfb772324393] [ 663.202106] env[63372]: DEBUG nova.compute.utils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Binding failed for port 61938f95-e588-4bdf-b61a-0795bd8279e3, please check neutron logs for more information. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 663.202106] env[63372]: DEBUG oslo_concurrency.lockutils [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 31.076s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.202106] env[63372]: DEBUG nova.objects.instance [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63372) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 663.204687] env[63372]: DEBUG nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Build of instance ae86a977-0a9a-46e5-9072-cfb772324393 was re-scheduled: Binding failed for port 61938f95-e588-4bdf-b61a-0795bd8279e3, please check neutron logs for more information. {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 663.205242] env[63372]: DEBUG nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Unplugging VIFs for instance {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 663.205480] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Acquiring lock "refresh_cache-ae86a977-0a9a-46e5-9072-cfb772324393" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.205687] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Acquired lock "refresh_cache-ae86a977-0a9a-46e5-9072-cfb772324393" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.205874] env[63372]: DEBUG nova.network.neutron [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 663.288020] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023720, 'name': CreateVM_Task, 'duration_secs': 0.424497} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.290958] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 663.294627] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.294817] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.295181] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 663.299030] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fa9399e-6a23-4aac-84bb-422b53543616 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.304444] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1023721, 'name': Rename_Task, 'duration_secs': 0.207521} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.305254] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 663.305490] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a6a714ae-9d08-4602-938d-ddfb5c7a6bee {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.310937] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023722, 'name': ReconfigVM_Task, 'duration_secs': 0.450711} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.311199] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Waiting for the task: (returnval){ [ 663.311199] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52df80f9-13b8-be1a-5b8d-d4edd6917138" [ 663.311199] env[63372]: _type = "Task" [ 663.311199] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.311777] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Reconfigured VM instance instance-0000001b to attach disk [datastore2] 6d4502b7-b9df-4fef-8400-9ff07b820e2b/6d4502b7-b9df-4fef-8400-9ff07b820e2b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 663.312468] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ed12058-a1e1-4d02-b564-6cb253eb4c37 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.317670] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Waiting for the task: (returnval){ [ 663.317670] env[63372]: value = "task-1023723" [ 663.317670] env[63372]: _type = "Task" [ 663.317670] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.328316] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 663.328316] env[63372]: value = "task-1023724" [ 663.328316] env[63372]: _type = "Task" [ 663.328316] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.328532] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52df80f9-13b8-be1a-5b8d-d4edd6917138, 'name': SearchDatastore_Task, 'duration_secs': 0.010694} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.329243] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.329511] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 663.329707] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.339308] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a78d8f-4e86-c8fd-6ac6-2586ea487a4e, 'name': SearchDatastore_Task, 'duration_secs': 0.019573} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.339530] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1023723, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.340176] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.340453] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] bbba9f28-045d-41ab-8539-5b2968fe3d54/bbba9f28-045d-41ab-8539-5b2968fe3d54.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 663.340704] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.340888] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 663.341109] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-998db0c0-6f5e-4285-b506-1868b28adb33 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.346348] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9463aea-170e-45ce-bc75-469a1aaa39a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.348387] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023724, 'name': Rename_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.354840] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 663.354840] env[63372]: value = "task-1023725" [ 663.354840] env[63372]: _type = "Task" [ 663.354840] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.360306] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 663.360306] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 663.360480] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6263f0c1-76b0-4f3e-bef2-da564ec5fcd6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.365735] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023725, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.370532] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Waiting for the task: (returnval){ [ 663.370532] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521c6aed-c0a9-7134-4a52-ea901ac3a483" [ 663.370532] env[63372]: _type = "Task" [ 663.370532] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.377792] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521c6aed-c0a9-7134-4a52-ea901ac3a483, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.476577] env[63372]: DEBUG nova.network.neutron [-] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.572417] env[63372]: DEBUG nova.network.neutron [req-d543cdea-e68d-409e-861d-d5b75e75f81e req-6ad79ada-619f-459d-acec-bb7223e29cdf service nova] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Updated VIF entry in instance network info cache for port fd39de6f-7b40-4e7b-a88c-7996e802c774. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 663.572774] env[63372]: DEBUG nova.network.neutron [req-d543cdea-e68d-409e-861d-d5b75e75f81e req-6ad79ada-619f-459d-acec-bb7223e29cdf service nova] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Updating instance_info_cache with network_info: [{"id": "fd39de6f-7b40-4e7b-a88c-7996e802c774", "address": "fa:16:3e:21:de:13", "network": {"id": "07f988a8-80e0-4a49-a8c1-df19313a423c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-2017897237-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "baadac70098a43fab6dfb45f6a043548", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f85835c8-5d0c-4b2f-97c4-6c4006580f79", "external-id": "nsx-vlan-transportzone-245", "segmentation_id": 245, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd39de6f-7b", "ovs_interfaceid": "fd39de6f-7b40-4e7b-a88c-7996e802c774", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.742426] env[63372]: DEBUG nova.network.neutron [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.831593] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1023723, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.840702] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023724, 'name': Rename_Task, 'duration_secs': 0.181554} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.840984] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 663.845831] env[63372]: DEBUG nova.network.neutron [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Successfully updated port: 7b5b0928-bb42-4396-bf97-ad9cb3423a7a {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 663.851312] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8cc620cc-8cfb-40d5-b896-c93d33fc922d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.864599] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 663.864599] env[63372]: value = "task-1023726" [ 663.864599] env[63372]: _type = "Task" [ 663.864599] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.872834] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023725, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.881068] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023726, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.882130] env[63372]: DEBUG nova.network.neutron [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.889753] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521c6aed-c0a9-7134-4a52-ea901ac3a483, 'name': SearchDatastore_Task, 'duration_secs': 0.008407} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.890607] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcfe7481-00b3-4bd9-bb50-02ba348921f2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.896773] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Waiting for the task: (returnval){ [ 663.896773] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520ae331-ba9b-2e65-23e1-2caa8d0564af" [ 663.896773] env[63372]: _type = "Task" [ 663.896773] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.906419] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520ae331-ba9b-2e65-23e1-2caa8d0564af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.979560] env[63372]: INFO nova.compute.manager [-] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Took 1.42 seconds to deallocate network for instance. [ 664.076092] env[63372]: DEBUG oslo_concurrency.lockutils [req-d543cdea-e68d-409e-861d-d5b75e75f81e req-6ad79ada-619f-459d-acec-bb7223e29cdf service nova] Releasing lock "refresh_cache-9404a0f0-a3e5-43ee-bc20-e9566cfe5a60" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.218424] env[63372]: DEBUG oslo_concurrency.lockutils [None req-936cf0ac-3881-4874-8574-1362facd4125 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.219091] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.069s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.221266] env[63372]: INFO nova.compute.claims [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.231819] env[63372]: DEBUG nova.compute.manager [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Received event network-vif-deleted-8815dc60-cd0a-42fb-8e16-875baf0e4298 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 664.232029] env[63372]: DEBUG nova.compute.manager [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Received event network-vif-plugged-7b5b0928-bb42-4396-bf97-ad9cb3423a7a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 664.232219] env[63372]: DEBUG oslo_concurrency.lockutils [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] Acquiring lock "5adf6952-e019-4ea0-be91-8fe5fb83d53e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.232413] env[63372]: DEBUG oslo_concurrency.lockutils [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] Lock "5adf6952-e019-4ea0-be91-8fe5fb83d53e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.232567] env[63372]: DEBUG oslo_concurrency.lockutils [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] Lock "5adf6952-e019-4ea0-be91-8fe5fb83d53e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.232728] env[63372]: DEBUG nova.compute.manager [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] No waiting events found dispatching network-vif-plugged-7b5b0928-bb42-4396-bf97-ad9cb3423a7a {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 664.232881] env[63372]: WARNING nova.compute.manager [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Received unexpected event network-vif-plugged-7b5b0928-bb42-4396-bf97-ad9cb3423a7a for instance with vm_state building and task_state spawning. [ 664.234268] env[63372]: DEBUG nova.compute.manager [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Received event network-changed-7b5b0928-bb42-4396-bf97-ad9cb3423a7a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 664.234923] env[63372]: DEBUG nova.compute.manager [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Refreshing instance network info cache due to event network-changed-7b5b0928-bb42-4396-bf97-ad9cb3423a7a. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 664.234923] env[63372]: DEBUG oslo_concurrency.lockutils [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] Acquiring lock "refresh_cache-5adf6952-e019-4ea0-be91-8fe5fb83d53e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.234923] env[63372]: DEBUG oslo_concurrency.lockutils [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] Acquired lock "refresh_cache-5adf6952-e019-4ea0-be91-8fe5fb83d53e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.235162] env[63372]: DEBUG nova.network.neutron [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Refreshing network info cache for port 7b5b0928-bb42-4396-bf97-ad9cb3423a7a {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 664.332878] env[63372]: DEBUG oslo_vmware.api [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1023723, 'name': PowerOnVM_Task, 'duration_secs': 0.8132} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.333161] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 664.333355] env[63372]: INFO nova.compute.manager [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Took 8.77 seconds to spawn the instance on the hypervisor. [ 664.333528] env[63372]: DEBUG nova.compute.manager [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 664.334288] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf5a20ad-b265-4fb7-a053-b980c4680647 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.355171] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Acquiring lock "refresh_cache-5adf6952-e019-4ea0-be91-8fe5fb83d53e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.365339] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023725, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.616561} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.365722] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] bbba9f28-045d-41ab-8539-5b2968fe3d54/bbba9f28-045d-41ab-8539-5b2968fe3d54.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 664.365986] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 664.369051] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c7ce5595-b39e-4c98-af8e-a931d37f9ac2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.378336] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023726, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.379579] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 664.379579] env[63372]: value = "task-1023727" [ 664.379579] env[63372]: _type = "Task" [ 664.379579] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.387163] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Releasing lock "refresh_cache-ae86a977-0a9a-46e5-9072-cfb772324393" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.387372] env[63372]: DEBUG nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63372) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 664.387550] env[63372]: DEBUG nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.387713] env[63372]: DEBUG nova.network.neutron [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 664.389290] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023727, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.403320] env[63372]: DEBUG nova.network.neutron [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.410334] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520ae331-ba9b-2e65-23e1-2caa8d0564af, 'name': SearchDatastore_Task, 'duration_secs': 0.050419} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.411151] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.411417] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60/9404a0f0-a3e5-43ee-bc20-e9566cfe5a60.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 664.411673] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d12ff3b3-adb6-46db-bf30-f1b94b6b0d34 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.418124] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Waiting for the task: (returnval){ [ 664.418124] env[63372]: value = "task-1023728" [ 664.418124] env[63372]: _type = "Task" [ 664.418124] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.427592] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023728, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.487223] env[63372]: DEBUG oslo_concurrency.lockutils [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.769617] env[63372]: DEBUG nova.network.neutron [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.853309] env[63372]: DEBUG nova.network.neutron [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.854840] env[63372]: INFO nova.compute.manager [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Took 42.68 seconds to build instance. [ 664.881193] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023726, 'name': PowerOnVM_Task} progress is 71%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.888895] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023727, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081517} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.889905] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 664.890040] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb6328b-39ad-4a54-9da9-758ac0e4a52f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.918740] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Reconfiguring VM instance instance-0000001e to attach disk [datastore2] bbba9f28-045d-41ab-8539-5b2968fe3d54/bbba9f28-045d-41ab-8539-5b2968fe3d54.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 664.919966] env[63372]: DEBUG nova.network.neutron [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.921398] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ae789e6-81c1-47b6-a041-7ecb8050c426 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.945499] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023728, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.946805] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 664.946805] env[63372]: value = "task-1023729" [ 664.946805] env[63372]: _type = "Task" [ 664.946805] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.956319] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023729, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.357736] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2131a87d-5214-4b5c-8bda-f056cda8deec tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Lock "cf673ac1-2c7d-468b-83ec-c723d5182457" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.925s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.358354] env[63372]: DEBUG oslo_concurrency.lockutils [req-5a18707b-dd07-4810-8600-bc39d4fb901a req-e3553c14-72bf-4806-87be-e93405e9f6ba service nova] Releasing lock "refresh_cache-5adf6952-e019-4ea0-be91-8fe5fb83d53e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.359837] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Acquired lock "refresh_cache-5adf6952-e019-4ea0-be91-8fe5fb83d53e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.360040] env[63372]: DEBUG nova.network.neutron [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 665.384614] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023726, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.435546] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023728, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.440597] env[63372]: INFO nova.compute.manager [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] [instance: ae86a977-0a9a-46e5-9072-cfb772324393] Took 1.05 seconds to deallocate network for instance. [ 665.468482] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023729, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.798104] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38add539-f43a-4288-8eb6-08168b03b568 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.808040] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be6f5335-13ef-4fdc-b187-511cb51bd5d0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.834658] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c8aaba-5859-499e-b4cf-c0868e5cb0fa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.841960] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e6e61e1-51ae-42b7-a4a9-17304878ba0b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.854878] env[63372]: DEBUG nova.compute.provider_tree [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.864281] env[63372]: DEBUG nova.compute.manager [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 665.876890] env[63372]: DEBUG oslo_vmware.api [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023726, 'name': PowerOnVM_Task, 'duration_secs': 1.60162} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.877059] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 665.877258] env[63372]: INFO nova.compute.manager [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Took 15.31 seconds to spawn the instance on the hypervisor. [ 665.877430] env[63372]: DEBUG nova.compute.manager [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 665.878199] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb17e82a-aa19-4d8b-9f32-d5d98216652a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.899408] env[63372]: DEBUG nova.network.neutron [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.926029] env[63372]: DEBUG nova.compute.manager [req-23398e43-9bb5-4a2a-bb1f-ef88d3286d5d req-985163d1-e1b0-49a6-8f67-db7c1307ffc7 service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Received event network-changed-e56add8e-4342-4ab4-b2f2-c03f257f3d3a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 665.926029] env[63372]: DEBUG nova.compute.manager [req-23398e43-9bb5-4a2a-bb1f-ef88d3286d5d req-985163d1-e1b0-49a6-8f67-db7c1307ffc7 service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Refreshing instance network info cache due to event network-changed-e56add8e-4342-4ab4-b2f2-c03f257f3d3a. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 665.926029] env[63372]: DEBUG oslo_concurrency.lockutils [req-23398e43-9bb5-4a2a-bb1f-ef88d3286d5d req-985163d1-e1b0-49a6-8f67-db7c1307ffc7 service nova] Acquiring lock "refresh_cache-cf673ac1-2c7d-468b-83ec-c723d5182457" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.926029] env[63372]: DEBUG oslo_concurrency.lockutils [req-23398e43-9bb5-4a2a-bb1f-ef88d3286d5d req-985163d1-e1b0-49a6-8f67-db7c1307ffc7 service nova] Acquired lock "refresh_cache-cf673ac1-2c7d-468b-83ec-c723d5182457" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.926029] env[63372]: DEBUG nova.network.neutron [req-23398e43-9bb5-4a2a-bb1f-ef88d3286d5d req-985163d1-e1b0-49a6-8f67-db7c1307ffc7 service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Refreshing network info cache for port e56add8e-4342-4ab4-b2f2-c03f257f3d3a {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 665.935366] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023728, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.053026} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.935603] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60/9404a0f0-a3e5-43ee-bc20-e9566cfe5a60.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 665.935831] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 665.936079] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fa90ff5b-8089-4c10-89d2-a9b3cccd5cad {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.943535] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Waiting for the task: (returnval){ [ 665.943535] env[63372]: value = "task-1023730" [ 665.943535] env[63372]: _type = "Task" [ 665.943535] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.963026] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023730, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.965911] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023729, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.056977] env[63372]: DEBUG nova.network.neutron [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Updating instance_info_cache with network_info: [{"id": "7b5b0928-bb42-4396-bf97-ad9cb3423a7a", "address": "fa:16:3e:11:2d:37", "network": {"id": "75e9f57f-9426-4fdc-9cec-e8e228ad9b34", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-33678262-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8b350f69f6d43ec869511e0e8fdb98c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51ae336c-12cf-406a-b1ca-54e9ce553b3e", "external-id": "nsx-vlan-transportzone-30", "segmentation_id": 30, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b5b0928-bb", "ovs_interfaceid": "7b5b0928-bb42-4396-bf97-ad9cb3423a7a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.357585] env[63372]: DEBUG nova.scheduler.client.report [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.387541] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.396778] env[63372]: INFO nova.compute.manager [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Took 45.67 seconds to build instance. [ 666.456516] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023730, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066162} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.459982] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 666.463010] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-316fc344-32f2-4264-a987-6df4b8f85c6c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.470685] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023729, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.488864] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Reconfiguring VM instance instance-0000001f to attach disk [datastore2] 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60/9404a0f0-a3e5-43ee-bc20-e9566cfe5a60.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 666.489939] env[63372]: INFO nova.scheduler.client.report [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Deleted allocations for instance ae86a977-0a9a-46e5-9072-cfb772324393 [ 666.498248] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f4b7ebd-74fa-40fb-a63b-48c790dfa184 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.519574] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Waiting for the task: (returnval){ [ 666.519574] env[63372]: value = "task-1023731" [ 666.519574] env[63372]: _type = "Task" [ 666.519574] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.527887] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023731, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.559564] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Releasing lock "refresh_cache-5adf6952-e019-4ea0-be91-8fe5fb83d53e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.560666] env[63372]: DEBUG nova.compute.manager [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Instance network_info: |[{"id": "7b5b0928-bb42-4396-bf97-ad9cb3423a7a", "address": "fa:16:3e:11:2d:37", "network": {"id": "75e9f57f-9426-4fdc-9cec-e8e228ad9b34", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-33678262-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8b350f69f6d43ec869511e0e8fdb98c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51ae336c-12cf-406a-b1ca-54e9ce553b3e", "external-id": "nsx-vlan-transportzone-30", "segmentation_id": 30, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b5b0928-bb", "ovs_interfaceid": "7b5b0928-bb42-4396-bf97-ad9cb3423a7a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 666.560666] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:2d:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '51ae336c-12cf-406a-b1ca-54e9ce553b3e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7b5b0928-bb42-4396-bf97-ad9cb3423a7a', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 666.569191] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Creating folder: Project (c8b350f69f6d43ec869511e0e8fdb98c). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 666.572199] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9e9711d2-f6bd-4ac6-8114-84fac7fac211 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.583689] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Created folder: Project (c8b350f69f6d43ec869511e0e8fdb98c) in parent group-v227230. [ 666.583944] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Creating folder: Instances. Parent ref: group-v227258. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 666.584219] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d2c5e810-d07e-47c6-ba71-a5b9a3f4c253 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.599823] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Created folder: Instances in parent group-v227258. [ 666.600147] env[63372]: DEBUG oslo.service.loopingcall [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 666.601814] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 666.604568] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7a026dab-1fa6-42c3-9f26-24de7215362e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.627568] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 666.627568] env[63372]: value = "task-1023734" [ 666.627568] env[63372]: _type = "Task" [ 666.627568] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.636094] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023734, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.726464] env[63372]: DEBUG nova.network.neutron [req-23398e43-9bb5-4a2a-bb1f-ef88d3286d5d req-985163d1-e1b0-49a6-8f67-db7c1307ffc7 service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Updated VIF entry in instance network info cache for port e56add8e-4342-4ab4-b2f2-c03f257f3d3a. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 666.726844] env[63372]: DEBUG nova.network.neutron [req-23398e43-9bb5-4a2a-bb1f-ef88d3286d5d req-985163d1-e1b0-49a6-8f67-db7c1307ffc7 service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Updating instance_info_cache with network_info: [{"id": "e56add8e-4342-4ab4-b2f2-c03f257f3d3a", "address": "fa:16:3e:11:78:8c", "network": {"id": "79050166-5dc7-4ef3-9819-ac534e041985", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-2103555585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06cecd1557de407697ad403a85138481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24727047-6358-4015-86c1-394ab07fb88f", "external-id": "nsx-vlan-transportzone-476", "segmentation_id": 476, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape56add8e-43", "ovs_interfaceid": "e56add8e-4342-4ab4-b2f2-c03f257f3d3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.863290] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.644s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.864103] env[63372]: DEBUG nova.compute.manager [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 666.867101] env[63372]: DEBUG oslo_concurrency.lockutils [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.777s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.868213] env[63372]: DEBUG nova.objects.instance [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Lazy-loading 'resources' on Instance uuid 8986e782-4d63-4491-8100-4341a3149812 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 666.899166] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9a5509e2-bd9b-4b0f-883e-737196f129cc tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.034s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.963202] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023729, 'name': ReconfigVM_Task, 'duration_secs': 1.677737} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.963549] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Reconfigured VM instance instance-0000001e to attach disk [datastore2] bbba9f28-045d-41ab-8539-5b2968fe3d54/bbba9f28-045d-41ab-8539-5b2968fe3d54.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 666.964279] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-061ad217-5aff-41f5-a52e-74d7e96f6803 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.972016] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 666.972016] env[63372]: value = "task-1023735" [ 666.972016] env[63372]: _type = "Task" [ 666.972016] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.978970] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023735, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.015438] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5a5fae2-8499-4e4c-8712-0215958b6f8d tempest-ServersTestFqdnHostnames-171841681 tempest-ServersTestFqdnHostnames-171841681-project-member] Lock "ae86a977-0a9a-46e5-9072-cfb772324393" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.322s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.029411] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023731, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.124263] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.124564] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.124792] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.124975] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.125161] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.127600] env[63372]: INFO nova.compute.manager [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Terminating instance [ 667.133681] env[63372]: DEBUG nova.compute.manager [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 667.133838] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 667.134723] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5689b631-d1e0-4bee-a749-70cfda24603a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.146443] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023734, 'name': CreateVM_Task, 'duration_secs': 0.350587} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.147374] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 667.147712] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 667.148483] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.148661] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.149012] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 667.149268] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99877466-e72b-4132-8a57-0c3e1c771023 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.151012] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b7b9d5e-ea5a-4bc1-a982-bd87081d3ec1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.158486] env[63372]: DEBUG oslo_vmware.api [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 667.158486] env[63372]: value = "task-1023736" [ 667.158486] env[63372]: _type = "Task" [ 667.158486] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.163428] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Waiting for the task: (returnval){ [ 667.163428] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e4b8c8-9861-1a73-a6b1-449a402749b3" [ 667.163428] env[63372]: _type = "Task" [ 667.163428] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.169931] env[63372]: DEBUG oslo_vmware.api [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023736, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.175180] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e4b8c8-9861-1a73-a6b1-449a402749b3, 'name': SearchDatastore_Task, 'duration_secs': 0.010475} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.175492] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.175750] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 667.176040] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.176213] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.176398] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 667.176671] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a81ee6fc-7e50-4ced-9d31-756a6ed42a88 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.185356] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 667.185567] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 667.186343] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-819c2324-c702-4c46-9248-e959b365cb2d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.191968] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Waiting for the task: (returnval){ [ 667.191968] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5285093a-3118-a00e-7809-04b024a45f63" [ 667.191968] env[63372]: _type = "Task" [ 667.191968] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.199790] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5285093a-3118-a00e-7809-04b024a45f63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.229817] env[63372]: DEBUG oslo_concurrency.lockutils [req-23398e43-9bb5-4a2a-bb1f-ef88d3286d5d req-985163d1-e1b0-49a6-8f67-db7c1307ffc7 service nova] Releasing lock "refresh_cache-cf673ac1-2c7d-468b-83ec-c723d5182457" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.372124] env[63372]: DEBUG nova.compute.utils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 667.376994] env[63372]: DEBUG nova.compute.manager [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 667.377180] env[63372]: DEBUG nova.network.neutron [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 667.403664] env[63372]: DEBUG nova.compute.manager [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 667.435461] env[63372]: DEBUG nova.policy [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5578d2e3e8dd498eb60b456f282d0449', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7c0777c24e2e4db083de571e3a65fa15', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 667.480615] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023735, 'name': Rename_Task, 'duration_secs': 0.244463} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.480615] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 667.483634] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d48f8434-d96b-4ab2-9c6f-5db86981d7bc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.490519] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 667.490519] env[63372]: value = "task-1023737" [ 667.490519] env[63372]: _type = "Task" [ 667.490519] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.501409] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023737, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.518329] env[63372]: DEBUG nova.compute.manager [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 667.531317] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023731, 'name': ReconfigVM_Task, 'duration_secs': 0.670496} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.534844] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Reconfigured VM instance instance-0000001f to attach disk [datastore2] 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60/9404a0f0-a3e5-43ee-bc20-e9566cfe5a60.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 667.534844] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-584ba330-d9b1-4267-bb27-075dca369aaa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.541932] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Waiting for the task: (returnval){ [ 667.541932] env[63372]: value = "task-1023738" [ 667.541932] env[63372]: _type = "Task" [ 667.541932] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.552988] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023738, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.672014] env[63372]: DEBUG oslo_vmware.api [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023736, 'name': PowerOffVM_Task, 'duration_secs': 0.361555} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.672014] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 667.672014] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 667.672014] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05402b84-173e-4c87-a6ab-f3693c49618b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.702387] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5285093a-3118-a00e-7809-04b024a45f63, 'name': SearchDatastore_Task, 'duration_secs': 0.011704} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.705801] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03d50002-3208-44b7-891f-f17ffbbcb276 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.710734] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Waiting for the task: (returnval){ [ 667.710734] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5271eeb5-b9d5-ad4a-db57-8b4b675505f9" [ 667.710734] env[63372]: _type = "Task" [ 667.710734] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.718723] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5271eeb5-b9d5-ad4a-db57-8b4b675505f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.759386] env[63372]: DEBUG nova.network.neutron [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Successfully created port: 40415ba1-4a9b-45d8-90ff-d95ac4e3980f {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 667.821228] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ca29f9-ab18-479d-b201-02d691b0de99 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.825299] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 667.825579] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 667.825848] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Deleting the datastore file [datastore2] 6d4502b7-b9df-4fef-8400-9ff07b820e2b {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 667.826578] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8bb868f-d921-4232-a467-1e32055750d7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.832135] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3bcb9d-39e1-4cae-8948-23c40e18609e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.837015] env[63372]: DEBUG oslo_vmware.api [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 667.837015] env[63372]: value = "task-1023740" [ 667.837015] env[63372]: _type = "Task" [ 667.837015] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.874022] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6925c39d-004c-4b3c-8dba-af3815f76569 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.878586] env[63372]: DEBUG nova.compute.manager [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 667.881352] env[63372]: DEBUG oslo_vmware.api [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023740, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.886539] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c4f7def-a096-4387-9059-a8939c070774 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.901362] env[63372]: DEBUG nova.compute.provider_tree [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.925888] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.002060] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023737, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.047725] env[63372]: DEBUG oslo_concurrency.lockutils [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.050596] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023738, 'name': Rename_Task, 'duration_secs': 0.1843} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.050874] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 668.051154] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb1fbc54-967d-4d74-b9f5-369d6a0faca1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.057036] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Waiting for the task: (returnval){ [ 668.057036] env[63372]: value = "task-1023741" [ 668.057036] env[63372]: _type = "Task" [ 668.057036] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.066512] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023741, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.221425] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5271eeb5-b9d5-ad4a-db57-8b4b675505f9, 'name': SearchDatastore_Task, 'duration_secs': 0.011393} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.221559] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.221968] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 5adf6952-e019-4ea0-be91-8fe5fb83d53e/5adf6952-e019-4ea0-be91-8fe5fb83d53e.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 668.222151] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-158b8f37-f7fc-4c94-be0d-6db63804865b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.228815] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Waiting for the task: (returnval){ [ 668.228815] env[63372]: value = "task-1023742" [ 668.228815] env[63372]: _type = "Task" [ 668.228815] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.235596] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023742, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.347216] env[63372]: DEBUG oslo_vmware.api [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1023740, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.265945} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.347472] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 668.347657] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 668.347863] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 668.348064] env[63372]: INFO nova.compute.manager [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Took 1.21 seconds to destroy the instance on the hypervisor. [ 668.348307] env[63372]: DEBUG oslo.service.loopingcall [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 668.348496] env[63372]: DEBUG nova.compute.manager [-] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 668.348589] env[63372]: DEBUG nova.network.neutron [-] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 668.404211] env[63372]: DEBUG nova.scheduler.client.report [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.507240] env[63372]: DEBUG oslo_vmware.api [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1023737, 'name': PowerOnVM_Task, 'duration_secs': 0.575429} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.507495] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 668.507687] env[63372]: INFO nova.compute.manager [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Took 10.45 seconds to spawn the instance on the hypervisor. [ 668.507855] env[63372]: DEBUG nova.compute.manager [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 668.508704] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-770c12fd-87da-40d1-8468-7ae3b80afef9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.569273] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023741, 'name': PowerOnVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.742228] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023742, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.898169] env[63372]: DEBUG nova.compute.manager [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 668.909177] env[63372]: DEBUG oslo_concurrency.lockutils [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.042s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.911444] env[63372]: DEBUG oslo_concurrency.lockutils [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 27.650s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.911633] env[63372]: DEBUG nova.objects.instance [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63372) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 668.926789] env[63372]: DEBUG nova.virt.hardware [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 668.927024] env[63372]: DEBUG nova.virt.hardware [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 668.927626] env[63372]: DEBUG nova.virt.hardware [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 668.927626] env[63372]: DEBUG nova.virt.hardware [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 668.927626] env[63372]: DEBUG nova.virt.hardware [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 668.927626] env[63372]: DEBUG nova.virt.hardware [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 668.933374] env[63372]: DEBUG nova.virt.hardware [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 668.933374] env[63372]: DEBUG nova.virt.hardware [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 668.933374] env[63372]: DEBUG nova.virt.hardware [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 668.933374] env[63372]: DEBUG nova.virt.hardware [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 668.933374] env[63372]: DEBUG nova.virt.hardware [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.933374] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940de530-ed47-44d6-8090-d15b6ee21564 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.939386] env[63372]: INFO nova.scheduler.client.report [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Deleted allocations for instance 8986e782-4d63-4491-8100-4341a3149812 [ 668.949407] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-526e3ee9-8d5b-4dab-85a0-d67a3d175d6e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.957290] env[63372]: DEBUG nova.compute.manager [req-97d2f19f-cd27-4178-995a-65ddf28de784 req-9cbb4a14-967b-4f35-9d2d-264460ae88f1 service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Received event network-vif-deleted-a52a68b2-6f6c-4b4d-b377-b79e60f46dac {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 668.957650] env[63372]: INFO nova.compute.manager [req-97d2f19f-cd27-4178-995a-65ddf28de784 req-9cbb4a14-967b-4f35-9d2d-264460ae88f1 service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Neutron deleted interface a52a68b2-6f6c-4b4d-b377-b79e60f46dac; detaching it from the instance and deleting it from the info cache [ 668.958057] env[63372]: DEBUG nova.network.neutron [req-97d2f19f-cd27-4178-995a-65ddf28de784 req-9cbb4a14-967b-4f35-9d2d-264460ae88f1 service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Updating instance_info_cache with network_info: [{"id": "7f75cf3b-acb2-456d-896e-25184e6811c4", "address": "fa:16:3e:f0:dc:5b", "network": {"id": "c5598ebc-5665-48b1-8d08-dcd4b12eb9f4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-945495745", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f75cf3b-ac", "ovs_interfaceid": "7f75cf3b-acb2-456d-896e-25184e6811c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d1e8b347-53e9-4aeb-82d6-f9e255d24442", "address": "fa:16:3e:28:04:2c", "network": {"id": "cd0bbfa3-9d50-4c94-b6ef-8d2e4506d27d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-203381185", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.167", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1e8b347-53", "ovs_interfaceid": "d1e8b347-53e9-4aeb-82d6-f9e255d24442", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.028383] env[63372]: INFO nova.compute.manager [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Took 44.87 seconds to build instance. [ 669.070566] env[63372]: DEBUG oslo_vmware.api [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023741, 'name': PowerOnVM_Task, 'duration_secs': 0.701968} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.070566] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 669.070566] env[63372]: INFO nova.compute.manager [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Took 8.53 seconds to spawn the instance on the hypervisor. [ 669.070566] env[63372]: DEBUG nova.compute.manager [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 669.070566] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-168435d3-f79f-4a65-b785-f37a720bc30d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.241712] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023742, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542456} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.241712] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 5adf6952-e019-4ea0-be91-8fe5fb83d53e/5adf6952-e019-4ea0-be91-8fe5fb83d53e.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 669.241712] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 669.241712] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ff488a58-b4a0-4758-8f7d-767a892f456f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.249021] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Waiting for the task: (returnval){ [ 669.249021] env[63372]: value = "task-1023743" [ 669.249021] env[63372]: _type = "Task" [ 669.249021] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.254611] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023743, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.453013] env[63372]: DEBUG oslo_concurrency.lockutils [None req-305e4de3-6b7d-4083-8b1f-4a862a2ec927 tempest-ServerDiagnosticsV248Test-671044190 tempest-ServerDiagnosticsV248Test-671044190-project-member] Lock "8986e782-4d63-4491-8100-4341a3149812" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.081s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.474293] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8602847d-4681-44e5-a440-c274a9b52502 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.482728] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e954a163-c8c9-4e98-a4f7-5b36919f2fe7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.511619] env[63372]: DEBUG nova.compute.manager [req-97d2f19f-cd27-4178-995a-65ddf28de784 req-9cbb4a14-967b-4f35-9d2d-264460ae88f1 service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Detach interface failed, port_id=a52a68b2-6f6c-4b4d-b377-b79e60f46dac, reason: Instance 6d4502b7-b9df-4fef-8400-9ff07b820e2b could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 669.530971] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f645439-6981-482a-bc52-a9b656ac5b81 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.682s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.588893] env[63372]: INFO nova.compute.manager [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Took 43.37 seconds to build instance. [ 669.630805] env[63372]: DEBUG nova.network.neutron [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Successfully updated port: 40415ba1-4a9b-45d8-90ff-d95ac4e3980f {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 669.762761] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023743, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074825} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.763181] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 669.764037] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-552defd8-a41f-48dc-ba91-7bdca50306f5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.788015] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] 5adf6952-e019-4ea0-be91-8fe5fb83d53e/5adf6952-e019-4ea0-be91-8fe5fb83d53e.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 669.788377] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b696a5e2-0a99-427d-b918-be310948c671 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.810267] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Waiting for the task: (returnval){ [ 669.810267] env[63372]: value = "task-1023744" [ 669.810267] env[63372]: _type = "Task" [ 669.810267] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.819098] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023744, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.920804] env[63372]: DEBUG oslo_concurrency.lockutils [None req-47736a6f-b3be-4c75-bd85-cb83d90e41d9 tempest-ServersAdmin275Test-885126451 tempest-ServersAdmin275Test-885126451-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.921993] env[63372]: DEBUG oslo_concurrency.lockutils [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.756s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.922258] env[63372]: DEBUG nova.objects.instance [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Lazy-loading 'resources' on Instance uuid 715408b8-0e76-4b61-a342-b168377cb288 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 669.998967] env[63372]: DEBUG nova.compute.manager [req-39981171-ae60-4a98-af20-27ae86f8b4f6 req-2ab28ce4-ea11-4f38-b0c9-f861eaea0d3e service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Received event network-changed-b4cf27a3-fc17-4ac5-b96a-3be058f71e03 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 669.999166] env[63372]: DEBUG nova.compute.manager [req-39981171-ae60-4a98-af20-27ae86f8b4f6 req-2ab28ce4-ea11-4f38-b0c9-f861eaea0d3e service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Refreshing instance network info cache due to event network-changed-b4cf27a3-fc17-4ac5-b96a-3be058f71e03. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 669.999372] env[63372]: DEBUG oslo_concurrency.lockutils [req-39981171-ae60-4a98-af20-27ae86f8b4f6 req-2ab28ce4-ea11-4f38-b0c9-f861eaea0d3e service nova] Acquiring lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.999508] env[63372]: DEBUG oslo_concurrency.lockutils [req-39981171-ae60-4a98-af20-27ae86f8b4f6 req-2ab28ce4-ea11-4f38-b0c9-f861eaea0d3e service nova] Acquired lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.999700] env[63372]: DEBUG nova.network.neutron [req-39981171-ae60-4a98-af20-27ae86f8b4f6 req-2ab28ce4-ea11-4f38-b0c9-f861eaea0d3e service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Refreshing network info cache for port b4cf27a3-fc17-4ac5-b96a-3be058f71e03 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 670.034113] env[63372]: DEBUG nova.compute.manager [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 670.056163] env[63372]: DEBUG nova.network.neutron [-] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.091499] env[63372]: DEBUG oslo_concurrency.lockutils [None req-848d0bd1-81f7-42d4-8616-e8b5c04b6695 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Lock "9404a0f0-a3e5-43ee-bc20-e9566cfe5a60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.654s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.136503] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "refresh_cache-ac90a156-be00-4f62-a76e-e08914531167" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.136503] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquired lock "refresh_cache-ac90a156-be00-4f62-a76e-e08914531167" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.136796] env[63372]: DEBUG nova.network.neutron [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 670.327258] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023744, 'name': ReconfigVM_Task, 'duration_secs': 0.291022} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.327258] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Reconfigured VM instance instance-00000020 to attach disk [datastore2] 5adf6952-e019-4ea0-be91-8fe5fb83d53e/5adf6952-e019-4ea0-be91-8fe5fb83d53e.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 670.327853] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7096efa-cf70-42a3-98e9-e69b4567fdaf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.338158] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Waiting for the task: (returnval){ [ 670.338158] env[63372]: value = "task-1023745" [ 670.338158] env[63372]: _type = "Task" [ 670.338158] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.352494] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023745, 'name': Rename_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.364219] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Acquiring lock "9404a0f0-a3e5-43ee-bc20-e9566cfe5a60" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.364475] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Lock "9404a0f0-a3e5-43ee-bc20-e9566cfe5a60" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.364688] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Acquiring lock "9404a0f0-a3e5-43ee-bc20-e9566cfe5a60-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.364858] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Lock "9404a0f0-a3e5-43ee-bc20-e9566cfe5a60-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.365398] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Lock "9404a0f0-a3e5-43ee-bc20-e9566cfe5a60-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.369854] env[63372]: INFO nova.compute.manager [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Terminating instance [ 670.371863] env[63372]: DEBUG nova.compute.manager [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 670.372606] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 670.373035] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86da4319-7b22-45c0-8b85-043656880f96 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.381514] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 670.381514] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-03fb9459-ae40-4f1f-a54f-d1ff36ee5e8e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.388594] env[63372]: DEBUG oslo_vmware.api [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Waiting for the task: (returnval){ [ 670.388594] env[63372]: value = "task-1023746" [ 670.388594] env[63372]: _type = "Task" [ 670.388594] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.396757] env[63372]: DEBUG oslo_vmware.api [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023746, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.562447] env[63372]: INFO nova.compute.manager [-] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Took 2.21 seconds to deallocate network for instance. [ 670.565593] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.594158] env[63372]: DEBUG nova.compute.manager [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 670.709964] env[63372]: DEBUG nova.network.neutron [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.855435] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023745, 'name': Rename_Task, 'duration_secs': 0.142412} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.858211] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 670.859291] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fd69d87b-182c-474b-ad59-de143f5ba67b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.861891] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392165b2-1311-4596-b78e-cbed64fbfba7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.874668] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e58627-de9c-4efe-8c15-e5f4b354a2ad {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.878315] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Waiting for the task: (returnval){ [ 670.878315] env[63372]: value = "task-1023747" [ 670.878315] env[63372]: _type = "Task" [ 670.878315] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.923113] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8791e75-48ee-4520-8a24-80b6843e2cf4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.935552] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023747, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.938211] env[63372]: DEBUG oslo_vmware.api [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023746, 'name': PowerOffVM_Task, 'duration_secs': 0.193687} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.940514] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 670.940684] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 670.940992] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5fc7ae6f-fb73-487c-90aa-41c7ee3113a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.943756] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-971cce57-44dc-4420-be1f-23d4ca09c6d1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.958641] env[63372]: DEBUG nova.compute.provider_tree [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 670.997328] env[63372]: DEBUG nova.compute.manager [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Received event network-vif-deleted-d1e8b347-53e9-4aeb-82d6-f9e255d24442 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 670.997432] env[63372]: DEBUG nova.compute.manager [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Received event network-vif-plugged-40415ba1-4a9b-45d8-90ff-d95ac4e3980f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 670.997648] env[63372]: DEBUG oslo_concurrency.lockutils [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] Acquiring lock "ac90a156-be00-4f62-a76e-e08914531167-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.997862] env[63372]: DEBUG oslo_concurrency.lockutils [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] Lock "ac90a156-be00-4f62-a76e-e08914531167-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.998058] env[63372]: DEBUG oslo_concurrency.lockutils [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] Lock "ac90a156-be00-4f62-a76e-e08914531167-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.998226] env[63372]: DEBUG nova.compute.manager [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] No waiting events found dispatching network-vif-plugged-40415ba1-4a9b-45d8-90ff-d95ac4e3980f {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 670.998385] env[63372]: WARNING nova.compute.manager [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Received unexpected event network-vif-plugged-40415ba1-4a9b-45d8-90ff-d95ac4e3980f for instance with vm_state building and task_state spawning. [ 670.998540] env[63372]: DEBUG nova.compute.manager [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Received event network-vif-deleted-7f75cf3b-acb2-456d-896e-25184e6811c4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 670.998697] env[63372]: DEBUG nova.compute.manager [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Received event network-changed-40415ba1-4a9b-45d8-90ff-d95ac4e3980f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 670.998841] env[63372]: DEBUG nova.compute.manager [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Refreshing instance network info cache due to event network-changed-40415ba1-4a9b-45d8-90ff-d95ac4e3980f. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 670.999015] env[63372]: DEBUG oslo_concurrency.lockutils [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] Acquiring lock "refresh_cache-ac90a156-be00-4f62-a76e-e08914531167" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.005583] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 671.005583] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 671.005717] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Deleting the datastore file [datastore2] 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 671.005946] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f40a7c6-19ad-4022-9a9e-46f20e8814ed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.012084] env[63372]: DEBUG oslo_vmware.api [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Waiting for the task: (returnval){ [ 671.012084] env[63372]: value = "task-1023749" [ 671.012084] env[63372]: _type = "Task" [ 671.012084] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.020645] env[63372]: DEBUG oslo_vmware.api [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023749, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.021578] env[63372]: DEBUG nova.network.neutron [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Updating instance_info_cache with network_info: [{"id": "40415ba1-4a9b-45d8-90ff-d95ac4e3980f", "address": "fa:16:3e:f6:4e:2b", "network": {"id": "d4022663-3d99-4ea6-b2c8-7bf206d3a801", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2038051617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c0777c24e2e4db083de571e3a65fa15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ee018eb-75be-4037-a80a-07034d4eae35", "external-id": "nsx-vlan-transportzone-8", "segmentation_id": 8, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40415ba1-4a", "ovs_interfaceid": "40415ba1-4a9b-45d8-90ff-d95ac4e3980f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.074510] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.077583] env[63372]: DEBUG nova.network.neutron [req-39981171-ae60-4a98-af20-27ae86f8b4f6 req-2ab28ce4-ea11-4f38-b0c9-f861eaea0d3e service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Updated VIF entry in instance network info cache for port b4cf27a3-fc17-4ac5-b96a-3be058f71e03. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 671.077925] env[63372]: DEBUG nova.network.neutron [req-39981171-ae60-4a98-af20-27ae86f8b4f6 req-2ab28ce4-ea11-4f38-b0c9-f861eaea0d3e service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Updating instance_info_cache with network_info: [{"id": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "address": "fa:16:3e:17:c8:2e", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4cf27a3-fc", "ovs_interfaceid": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.125475] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.391024] env[63372]: DEBUG oslo_vmware.api [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023747, 'name': PowerOnVM_Task, 'duration_secs': 0.518115} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.391024] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 671.391024] env[63372]: INFO nova.compute.manager [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Took 8.28 seconds to spawn the instance on the hypervisor. [ 671.391024] env[63372]: DEBUG nova.compute.manager [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 671.391250] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf63ee51-8b96-4887-8694-333136c2d92b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.480344] env[63372]: ERROR nova.scheduler.client.report [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] [req-48b22cef-a3c6-470a-a5c7-8e80cbd5f633] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-48b22cef-a3c6-470a-a5c7-8e80cbd5f633"}]} [ 671.498995] env[63372]: DEBUG nova.scheduler.client.report [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Refreshing inventories for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 671.514873] env[63372]: DEBUG nova.scheduler.client.report [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Updating ProviderTree inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 671.515116] env[63372]: DEBUG nova.compute.provider_tree [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 671.525476] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Releasing lock "refresh_cache-ac90a156-be00-4f62-a76e-e08914531167" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.526145] env[63372]: DEBUG nova.compute.manager [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Instance network_info: |[{"id": "40415ba1-4a9b-45d8-90ff-d95ac4e3980f", "address": "fa:16:3e:f6:4e:2b", "network": {"id": "d4022663-3d99-4ea6-b2c8-7bf206d3a801", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2038051617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c0777c24e2e4db083de571e3a65fa15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ee018eb-75be-4037-a80a-07034d4eae35", "external-id": "nsx-vlan-transportzone-8", "segmentation_id": 8, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40415ba1-4a", "ovs_interfaceid": "40415ba1-4a9b-45d8-90ff-d95ac4e3980f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 671.529746] env[63372]: DEBUG oslo_concurrency.lockutils [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] Acquired lock "refresh_cache-ac90a156-be00-4f62-a76e-e08914531167" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.530372] env[63372]: DEBUG nova.network.neutron [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Refreshing network info cache for port 40415ba1-4a9b-45d8-90ff-d95ac4e3980f {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 671.531770] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:4e:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ee018eb-75be-4037-a80a-07034d4eae35', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '40415ba1-4a9b-45d8-90ff-d95ac4e3980f', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 671.540661] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Creating folder: Project (7c0777c24e2e4db083de571e3a65fa15). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 671.541009] env[63372]: DEBUG oslo_vmware.api [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Task: {'id': task-1023749, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204439} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.543047] env[63372]: DEBUG nova.scheduler.client.report [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Refreshing aggregate associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, aggregates: None {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 671.545016] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4dc32241-4e91-49bb-8ba5-c584be3dbdc3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.547247] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 671.547436] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 671.547616] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 671.547784] env[63372]: INFO nova.compute.manager [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Took 1.18 seconds to destroy the instance on the hypervisor. [ 671.548036] env[63372]: DEBUG oslo.service.loopingcall [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 671.552396] env[63372]: DEBUG nova.compute.manager [-] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.552396] env[63372]: DEBUG nova.network.neutron [-] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 671.561021] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Created folder: Project (7c0777c24e2e4db083de571e3a65fa15) in parent group-v227230. [ 671.561021] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Creating folder: Instances. Parent ref: group-v227261. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 671.561021] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4cc80ab5-5948-4733-b555-49dff3bf6281 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.569179] env[63372]: DEBUG nova.scheduler.client.report [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Refreshing trait associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 671.570898] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Created folder: Instances in parent group-v227261. [ 671.571331] env[63372]: DEBUG oslo.service.loopingcall [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 671.571808] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac90a156-be00-4f62-a76e-e08914531167] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 671.574180] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-784cbaf7-9d13-4e89-9715-a9fee62c68ac {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.589240] env[63372]: DEBUG oslo_concurrency.lockutils [req-39981171-ae60-4a98-af20-27ae86f8b4f6 req-2ab28ce4-ea11-4f38-b0c9-f861eaea0d3e service nova] Releasing lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.599966] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 671.599966] env[63372]: value = "task-1023752" [ 671.599966] env[63372]: _type = "Task" [ 671.599966] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.611653] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023752, 'name': CreateVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.912398] env[63372]: INFO nova.compute.manager [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Took 43.70 seconds to build instance. [ 671.976360] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9cdb7d-be17-47c9-bffe-c07351abbbb7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.985370] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54ceead-e3b5-4b57-953d-5ef48fcd186f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.020390] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e31ec35-93b6-435d-b58d-de177edf8848 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.028916] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b307f4eb-dad2-4aec-9743-1c3c9e842c9a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.046213] env[63372]: DEBUG nova.compute.provider_tree [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 672.107911] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023752, 'name': CreateVM_Task, 'duration_secs': 0.356149} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.107911] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac90a156-be00-4f62-a76e-e08914531167] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 672.108277] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.108586] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.108863] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 672.109120] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5507ed93-02a2-4566-8932-5e6ee8bd44ab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.114138] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 672.114138] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52eca67c-dd56-4630-ec7c-64354b9370d1" [ 672.114138] env[63372]: _type = "Task" [ 672.114138] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.122080] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52eca67c-dd56-4630-ec7c-64354b9370d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.289420] env[63372]: DEBUG nova.network.neutron [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Updated VIF entry in instance network info cache for port 40415ba1-4a9b-45d8-90ff-d95ac4e3980f. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 672.289861] env[63372]: DEBUG nova.network.neutron [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Updating instance_info_cache with network_info: [{"id": "40415ba1-4a9b-45d8-90ff-d95ac4e3980f", "address": "fa:16:3e:f6:4e:2b", "network": {"id": "d4022663-3d99-4ea6-b2c8-7bf206d3a801", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2038051617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c0777c24e2e4db083de571e3a65fa15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ee018eb-75be-4037-a80a-07034d4eae35", "external-id": "nsx-vlan-transportzone-8", "segmentation_id": 8, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40415ba1-4a", "ovs_interfaceid": "40415ba1-4a9b-45d8-90ff-d95ac4e3980f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.414647] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84cc4280-0052-4584-81fe-23825a033f20 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Lock "5adf6952-e019-4ea0-be91-8fe5fb83d53e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.950s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.485067] env[63372]: DEBUG nova.network.neutron [-] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.583026] env[63372]: DEBUG nova.scheduler.client.report [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 58 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 672.583309] env[63372]: DEBUG nova.compute.provider_tree [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 58 to 59 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 672.583658] env[63372]: DEBUG nova.compute.provider_tree [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 672.628640] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52eca67c-dd56-4630-ec7c-64354b9370d1, 'name': SearchDatastore_Task, 'duration_secs': 0.021259} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.628975] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.629298] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 672.629536] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.629714] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.629958] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 672.630108] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9182a939-443f-41e2-81bc-363643278d0c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.640758] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 672.642110] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 672.642110] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d6e94e5-92b0-4202-8585-76a7da1c92ff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.648153] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 672.648153] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5232259a-b0af-eb45-63c6-76561f3ec4c2" [ 672.648153] env[63372]: _type = "Task" [ 672.648153] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.658383] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5232259a-b0af-eb45-63c6-76561f3ec4c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.794700] env[63372]: DEBUG oslo_concurrency.lockutils [req-187935b2-4c87-4341-802d-f313a87adcf6 req-8eae29bf-66cc-43b5-a30f-04c91602f026 service nova] Releasing lock "refresh_cache-ac90a156-be00-4f62-a76e-e08914531167" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.919041] env[63372]: DEBUG nova.compute.manager [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 672.989726] env[63372]: INFO nova.compute.manager [-] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Took 1.44 seconds to deallocate network for instance. [ 673.034605] env[63372]: DEBUG nova.compute.manager [req-32ae1a67-65d3-4e71-9710-daa14561a71e req-fd58a369-2c11-414e-8620-a4337c844df9 service nova] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Received event network-vif-deleted-fd39de6f-7b40-4e7b-a88c-7996e802c774 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 673.058966] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Acquiring lock "5adf6952-e019-4ea0-be91-8fe5fb83d53e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.058966] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Lock "5adf6952-e019-4ea0-be91-8fe5fb83d53e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.058966] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Acquiring lock "5adf6952-e019-4ea0-be91-8fe5fb83d53e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.058966] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Lock "5adf6952-e019-4ea0-be91-8fe5fb83d53e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.059258] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Lock "5adf6952-e019-4ea0-be91-8fe5fb83d53e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.062627] env[63372]: INFO nova.compute.manager [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Terminating instance [ 673.065622] env[63372]: DEBUG nova.compute.manager [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 673.065622] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 673.066616] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f3b47d-1deb-46fd-93b0-eecc3896f704 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.074634] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 673.074950] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-30491ec9-ceed-40ed-9cff-3add214848dd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.082191] env[63372]: DEBUG oslo_vmware.api [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Waiting for the task: (returnval){ [ 673.082191] env[63372]: value = "task-1023753" [ 673.082191] env[63372]: _type = "Task" [ 673.082191] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.090716] env[63372]: DEBUG oslo_concurrency.lockutils [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.169s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.092806] env[63372]: DEBUG oslo_vmware.api [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023753, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.093363] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.197s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.094959] env[63372]: INFO nova.compute.claims [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 673.127493] env[63372]: INFO nova.scheduler.client.report [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Deleted allocations for instance 715408b8-0e76-4b61-a342-b168377cb288 [ 673.164817] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5232259a-b0af-eb45-63c6-76561f3ec4c2, 'name': SearchDatastore_Task, 'duration_secs': 0.029213} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.165995] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce702dce-6a39-4726-b110-0c83fbc82552 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.173995] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 673.173995] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52c63112-4f92-97f1-d233-a6c80fc7b51e" [ 673.173995] env[63372]: _type = "Task" [ 673.173995] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.183550] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c63112-4f92-97f1-d233-a6c80fc7b51e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.447884] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.497416] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.598909] env[63372]: DEBUG oslo_vmware.api [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023753, 'name': PowerOffVM_Task, 'duration_secs': 0.23083} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.601210] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 673.605019] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 673.605019] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1151c55-95e9-4570-b24d-1ced176e7872 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.643484] env[63372]: DEBUG oslo_concurrency.lockutils [None req-aa77224c-f11a-4936-bce2-943ac862aef0 tempest-ServersAdmin275Test-2002679970 tempest-ServersAdmin275Test-2002679970-project-member] Lock "715408b8-0e76-4b61-a342-b168377cb288" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.170s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.672024] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 673.672024] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 673.672024] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Deleting the datastore file [datastore2] 5adf6952-e019-4ea0-be91-8fe5fb83d53e {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 673.672024] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6120e9d9-4f21-4acc-ae53-f824c4249238 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.679654] env[63372]: DEBUG oslo_vmware.api [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Waiting for the task: (returnval){ [ 673.679654] env[63372]: value = "task-1023755" [ 673.679654] env[63372]: _type = "Task" [ 673.679654] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.692037] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c63112-4f92-97f1-d233-a6c80fc7b51e, 'name': SearchDatastore_Task, 'duration_secs': 0.012324} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.692037] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.692037] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] ac90a156-be00-4f62-a76e-e08914531167/ac90a156-be00-4f62-a76e-e08914531167.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 673.692037] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c96f87a3-4ed0-4164-8ce7-48c31e8e35ed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.696850] env[63372]: DEBUG oslo_vmware.api [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023755, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.703592] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 673.703592] env[63372]: value = "task-1023756" [ 673.703592] env[63372]: _type = "Task" [ 673.703592] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.712377] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1023756, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.196446] env[63372]: DEBUG oslo_vmware.api [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Task: {'id': task-1023755, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167667} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.199265] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 674.199772] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 674.200316] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 674.200825] env[63372]: INFO nova.compute.manager [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 674.201224] env[63372]: DEBUG oslo.service.loopingcall [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 674.201924] env[63372]: DEBUG nova.compute.manager [-] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 674.202159] env[63372]: DEBUG nova.network.neutron [-] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 674.217094] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1023756, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501447} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.217094] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] ac90a156-be00-4f62-a76e-e08914531167/ac90a156-be00-4f62-a76e-e08914531167.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 674.217094] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 674.217094] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5543036-ffdd-4652-b3ad-729169cac62b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.224618] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 674.224618] env[63372]: value = "task-1023757" [ 674.224618] env[63372]: _type = "Task" [ 674.224618] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.237887] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1023757, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.510218] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637d06dc-fd14-41e9-98cf-1f27b46fea1b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.517819] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27925f5b-d3e9-4da5-9ed0-284b59b59da2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.553644] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8383d353-52f3-4d9d-ba69-b8ede59ac854 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.562122] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5709b5c-9490-42bd-a988-d594090ecfa9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.580269] env[63372]: DEBUG nova.compute.provider_tree [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.744020] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1023757, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076172} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.744020] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 674.744020] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f942a1a-1446-461e-aed8-c6927aaec336 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.772592] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] ac90a156-be00-4f62-a76e-e08914531167/ac90a156-be00-4f62-a76e-e08914531167.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 674.773104] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-999564e6-97c3-4b9b-bf12-49c65ef527cd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.800240] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 674.800240] env[63372]: value = "task-1023758" [ 674.800240] env[63372]: _type = "Task" [ 674.800240] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.807995] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1023758, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.985667] env[63372]: DEBUG nova.network.neutron [-] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.068056] env[63372]: DEBUG nova.compute.manager [req-48ae77cc-17ff-4697-94ee-73ed65dc2f8b req-f3757604-8869-4c7b-8b11-299c883749e9 service nova] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Received event network-vif-deleted-7b5b0928-bb42-4396-bf97-ad9cb3423a7a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 675.084995] env[63372]: DEBUG nova.scheduler.client.report [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.313035] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1023758, 'name': ReconfigVM_Task, 'duration_secs': 0.288485} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.313035] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Reconfigured VM instance instance-00000021 to attach disk [datastore2] ac90a156-be00-4f62-a76e-e08914531167/ac90a156-be00-4f62-a76e-e08914531167.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 675.313413] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cf2a053b-fd64-4059-ba94-6d2ca28c4e91 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.321766] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 675.321766] env[63372]: value = "task-1023759" [ 675.321766] env[63372]: _type = "Task" [ 675.321766] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.330222] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1023759, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.490387] env[63372]: INFO nova.compute.manager [-] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Took 1.29 seconds to deallocate network for instance. [ 675.591204] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.591291] env[63372]: DEBUG nova.compute.manager [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 675.594520] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.779s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.598074] env[63372]: INFO nova.compute.claims [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 675.833822] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1023759, 'name': Rename_Task, 'duration_secs': 0.147642} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.833822] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 675.833932] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-065e3051-0958-4529-979f-60dce4d62537 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.841063] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 675.841063] env[63372]: value = "task-1023760" [ 675.841063] env[63372]: _type = "Task" [ 675.841063] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.848769] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1023760, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.000090] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.096842] env[63372]: DEBUG nova.compute.utils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 676.102504] env[63372]: DEBUG nova.compute.manager [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 676.102688] env[63372]: DEBUG nova.network.neutron [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 676.191454] env[63372]: DEBUG nova.policy [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5039f0556904497c8cb69ab549d9fb32', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d4892a06e8b45e1873801404caf0739', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 676.356568] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1023760, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.572436] env[63372]: DEBUG nova.network.neutron [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Successfully created port: e018b1c1-a38b-4542-a2b4-bf87f12453ae {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 676.615299] env[63372]: DEBUG nova.compute.manager [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 676.860215] env[63372]: DEBUG oslo_vmware.api [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1023760, 'name': PowerOnVM_Task, 'duration_secs': 0.655365} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.860347] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 676.860623] env[63372]: INFO nova.compute.manager [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Took 7.96 seconds to spawn the instance on the hypervisor. [ 676.860727] env[63372]: DEBUG nova.compute.manager [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 676.864976] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ef43bf-5352-4f9c-ba23-25756f6da990 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.012801] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49bd8ff4-0a31-4dc2-adf6-9f89f97d56e3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.020024] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbabfd4a-9bb2-41dd-829e-c8a1b9dbfe90 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.051176] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45337f6c-8f7f-4ddb-bd94-b46a9dbdb30a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.059552] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdbd92af-bdc2-4d8a-b156-d901115feedd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.072959] env[63372]: DEBUG nova.compute.provider_tree [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 677.384958] env[63372]: INFO nova.compute.manager [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Took 44.26 seconds to build instance. [ 677.606849] env[63372]: DEBUG nova.scheduler.client.report [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 59 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 677.607183] env[63372]: DEBUG nova.compute.provider_tree [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 59 to 60 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 677.607378] env[63372]: DEBUG nova.compute.provider_tree [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 677.628572] env[63372]: DEBUG nova.compute.manager [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 677.657803] env[63372]: DEBUG nova.virt.hardware [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.658083] env[63372]: DEBUG nova.virt.hardware [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.658243] env[63372]: DEBUG nova.virt.hardware [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.658434] env[63372]: DEBUG nova.virt.hardware [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.658603] env[63372]: DEBUG nova.virt.hardware [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.658711] env[63372]: DEBUG nova.virt.hardware [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.658919] env[63372]: DEBUG nova.virt.hardware [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.659197] env[63372]: DEBUG nova.virt.hardware [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.659241] env[63372]: DEBUG nova.virt.hardware [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.659389] env[63372]: DEBUG nova.virt.hardware [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.659591] env[63372]: DEBUG nova.virt.hardware [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.660429] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2410e523-cd59-4358-8535-08d3b8fbacd4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.668419] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baaba98c-0855-4314-96fd-01064ef3d3e8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.887651] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6a92cfa6-5bd6-45da-ae79-4a5b496f0e35 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "ac90a156-be00-4f62-a76e-e08914531167" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.344s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.117026] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.117026] env[63372]: DEBUG nova.compute.manager [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 678.121501] env[63372]: DEBUG nova.network.neutron [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Successfully updated port: e018b1c1-a38b-4542-a2b4-bf87f12453ae {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 678.122575] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.952s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.124368] env[63372]: INFO nova.compute.claims [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 678.199571] env[63372]: DEBUG nova.compute.manager [req-c6708529-e198-4f06-89ec-6bbb05aa64e3 req-6b9db4da-1658-471b-adf3-53cd1ee45c30 service nova] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Received event network-vif-plugged-e018b1c1-a38b-4542-a2b4-bf87f12453ae {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 678.199737] env[63372]: DEBUG oslo_concurrency.lockutils [req-c6708529-e198-4f06-89ec-6bbb05aa64e3 req-6b9db4da-1658-471b-adf3-53cd1ee45c30 service nova] Acquiring lock "5819c38e-2cf9-4d16-b28a-5f23d35c3d44-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.199986] env[63372]: DEBUG oslo_concurrency.lockutils [req-c6708529-e198-4f06-89ec-6bbb05aa64e3 req-6b9db4da-1658-471b-adf3-53cd1ee45c30 service nova] Lock "5819c38e-2cf9-4d16-b28a-5f23d35c3d44-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.200175] env[63372]: DEBUG oslo_concurrency.lockutils [req-c6708529-e198-4f06-89ec-6bbb05aa64e3 req-6b9db4da-1658-471b-adf3-53cd1ee45c30 service nova] Lock "5819c38e-2cf9-4d16-b28a-5f23d35c3d44-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.201247] env[63372]: DEBUG nova.compute.manager [req-c6708529-e198-4f06-89ec-6bbb05aa64e3 req-6b9db4da-1658-471b-adf3-53cd1ee45c30 service nova] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] No waiting events found dispatching network-vif-plugged-e018b1c1-a38b-4542-a2b4-bf87f12453ae {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 678.201247] env[63372]: WARNING nova.compute.manager [req-c6708529-e198-4f06-89ec-6bbb05aa64e3 req-6b9db4da-1658-471b-adf3-53cd1ee45c30 service nova] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Received unexpected event network-vif-plugged-e018b1c1-a38b-4542-a2b4-bf87f12453ae for instance with vm_state building and task_state spawning. [ 678.392934] env[63372]: DEBUG nova.compute.manager [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 678.631289] env[63372]: DEBUG nova.compute.utils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 678.637608] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "refresh_cache-5819c38e-2cf9-4d16-b28a-5f23d35c3d44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.637833] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquired lock "refresh_cache-5819c38e-2cf9-4d16-b28a-5f23d35c3d44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.637987] env[63372]: DEBUG nova.network.neutron [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 678.639373] env[63372]: DEBUG nova.compute.manager [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 678.639373] env[63372]: DEBUG nova.network.neutron [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 678.719637] env[63372]: DEBUG nova.policy [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb233d4b6fe54fefae3046d62d167e34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d0c0f315749429dbae22a72ae82b1b2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 678.914743] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.988615] env[63372]: DEBUG nova.compute.manager [req-a145bdaf-e8c7-4094-ac2a-495bd34f599c req-696db557-5f4a-42c6-a7b2-aa9eb91c03c7 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Received event network-changed-40415ba1-4a9b-45d8-90ff-d95ac4e3980f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 678.988615] env[63372]: DEBUG nova.compute.manager [req-a145bdaf-e8c7-4094-ac2a-495bd34f599c req-696db557-5f4a-42c6-a7b2-aa9eb91c03c7 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Refreshing instance network info cache due to event network-changed-40415ba1-4a9b-45d8-90ff-d95ac4e3980f. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 678.988615] env[63372]: DEBUG oslo_concurrency.lockutils [req-a145bdaf-e8c7-4094-ac2a-495bd34f599c req-696db557-5f4a-42c6-a7b2-aa9eb91c03c7 service nova] Acquiring lock "refresh_cache-ac90a156-be00-4f62-a76e-e08914531167" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.988615] env[63372]: DEBUG oslo_concurrency.lockutils [req-a145bdaf-e8c7-4094-ac2a-495bd34f599c req-696db557-5f4a-42c6-a7b2-aa9eb91c03c7 service nova] Acquired lock "refresh_cache-ac90a156-be00-4f62-a76e-e08914531167" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.988910] env[63372]: DEBUG nova.network.neutron [req-a145bdaf-e8c7-4094-ac2a-495bd34f599c req-696db557-5f4a-42c6-a7b2-aa9eb91c03c7 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Refreshing network info cache for port 40415ba1-4a9b-45d8-90ff-d95ac4e3980f {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 679.080520] env[63372]: DEBUG nova.network.neutron [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Successfully created port: fff45461-b9dc-4d60-989d-c7460bc0becb {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 679.144653] env[63372]: DEBUG nova.compute.manager [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 679.209102] env[63372]: DEBUG nova.network.neutron [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.462468] env[63372]: DEBUG nova.network.neutron [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Updating instance_info_cache with network_info: [{"id": "e018b1c1-a38b-4542-a2b4-bf87f12453ae", "address": "fa:16:3e:cf:8c:50", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.85", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape018b1c1-a3", "ovs_interfaceid": "e018b1c1-a38b-4542-a2b4-bf87f12453ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.576035] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be72d13-77b4-4b30-9c1a-e64d77a546b3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.585495] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77897e8-20b1-4a89-b8de-c6aa1f1a833f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.621167] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c251113e-469d-4eac-9143-5bad01c293ce {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.631191] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c6a4b0-9021-4286-887e-8194bf17815c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.646169] env[63372]: DEBUG nova.compute.provider_tree [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.770443] env[63372]: DEBUG nova.network.neutron [req-a145bdaf-e8c7-4094-ac2a-495bd34f599c req-696db557-5f4a-42c6-a7b2-aa9eb91c03c7 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Updated VIF entry in instance network info cache for port 40415ba1-4a9b-45d8-90ff-d95ac4e3980f. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 679.771170] env[63372]: DEBUG nova.network.neutron [req-a145bdaf-e8c7-4094-ac2a-495bd34f599c req-696db557-5f4a-42c6-a7b2-aa9eb91c03c7 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Updating instance_info_cache with network_info: [{"id": "40415ba1-4a9b-45d8-90ff-d95ac4e3980f", "address": "fa:16:3e:f6:4e:2b", "network": {"id": "d4022663-3d99-4ea6-b2c8-7bf206d3a801", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2038051617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c0777c24e2e4db083de571e3a65fa15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ee018eb-75be-4037-a80a-07034d4eae35", "external-id": "nsx-vlan-transportzone-8", "segmentation_id": 8, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40415ba1-4a", "ovs_interfaceid": "40415ba1-4a9b-45d8-90ff-d95ac4e3980f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.965974] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Releasing lock "refresh_cache-5819c38e-2cf9-4d16-b28a-5f23d35c3d44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.965974] env[63372]: DEBUG nova.compute.manager [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Instance network_info: |[{"id": "e018b1c1-a38b-4542-a2b4-bf87f12453ae", "address": "fa:16:3e:cf:8c:50", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.85", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape018b1c1-a3", "ovs_interfaceid": "e018b1c1-a38b-4542-a2b4-bf87f12453ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 679.965974] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:8c:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e018b1c1-a38b-4542-a2b4-bf87f12453ae', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 679.973289] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Creating folder: Project (8d4892a06e8b45e1873801404caf0739). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 679.973578] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f5fbf2a7-62c6-48ef-bdac-adaf425eb6c8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.984043] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Created folder: Project (8d4892a06e8b45e1873801404caf0739) in parent group-v227230. [ 679.984307] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Creating folder: Instances. Parent ref: group-v227267. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 679.984588] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4911f701-5bf3-48dd-ac58-56ba3affc847 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.992261] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Created folder: Instances in parent group-v227267. [ 679.992496] env[63372]: DEBUG oslo.service.loopingcall [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.992683] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 679.992877] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b7c725c-624e-4772-82c2-46821b1523e3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.011117] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 680.011117] env[63372]: value = "task-1023767" [ 680.011117] env[63372]: _type = "Task" [ 680.011117] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.020960] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023767, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.149931] env[63372]: DEBUG nova.scheduler.client.report [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.161066] env[63372]: DEBUG nova.compute.manager [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 680.190776] env[63372]: DEBUG nova.virt.hardware [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 680.191040] env[63372]: DEBUG nova.virt.hardware [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 680.191204] env[63372]: DEBUG nova.virt.hardware [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 680.191386] env[63372]: DEBUG nova.virt.hardware [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 680.191530] env[63372]: DEBUG nova.virt.hardware [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 680.191671] env[63372]: DEBUG nova.virt.hardware [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 680.191871] env[63372]: DEBUG nova.virt.hardware [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 680.192042] env[63372]: DEBUG nova.virt.hardware [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 680.192209] env[63372]: DEBUG nova.virt.hardware [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 680.192365] env[63372]: DEBUG nova.virt.hardware [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 680.192533] env[63372]: DEBUG nova.virt.hardware [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 680.193401] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3fe54fc-5de5-4c20-b2c4-b7711c40b136 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.201424] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a507cd7-fd9d-4c28-9b21-276cc6916b6d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.228784] env[63372]: DEBUG nova.compute.manager [req-55254c9b-1b75-4b42-8a60-3d6b0e7ae2d7 req-5726b214-3ddd-412d-a2c2-1455233879ca service nova] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Received event network-changed-e018b1c1-a38b-4542-a2b4-bf87f12453ae {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 680.229083] env[63372]: DEBUG nova.compute.manager [req-55254c9b-1b75-4b42-8a60-3d6b0e7ae2d7 req-5726b214-3ddd-412d-a2c2-1455233879ca service nova] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Refreshing instance network info cache due to event network-changed-e018b1c1-a38b-4542-a2b4-bf87f12453ae. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 680.229366] env[63372]: DEBUG oslo_concurrency.lockutils [req-55254c9b-1b75-4b42-8a60-3d6b0e7ae2d7 req-5726b214-3ddd-412d-a2c2-1455233879ca service nova] Acquiring lock "refresh_cache-5819c38e-2cf9-4d16-b28a-5f23d35c3d44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.229559] env[63372]: DEBUG oslo_concurrency.lockutils [req-55254c9b-1b75-4b42-8a60-3d6b0e7ae2d7 req-5726b214-3ddd-412d-a2c2-1455233879ca service nova] Acquired lock "refresh_cache-5819c38e-2cf9-4d16-b28a-5f23d35c3d44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.229753] env[63372]: DEBUG nova.network.neutron [req-55254c9b-1b75-4b42-8a60-3d6b0e7ae2d7 req-5726b214-3ddd-412d-a2c2-1455233879ca service nova] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Refreshing network info cache for port e018b1c1-a38b-4542-a2b4-bf87f12453ae {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 680.275261] env[63372]: DEBUG oslo_concurrency.lockutils [req-a145bdaf-e8c7-4094-ac2a-495bd34f599c req-696db557-5f4a-42c6-a7b2-aa9eb91c03c7 service nova] Releasing lock "refresh_cache-ac90a156-be00-4f62-a76e-e08914531167" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.521268] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023767, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.657309] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.657828] env[63372]: DEBUG nova.compute.manager [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 680.660933] env[63372]: DEBUG oslo_concurrency.lockutils [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.174s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.660933] env[63372]: DEBUG nova.objects.instance [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Lazy-loading 'resources' on Instance uuid 779e3559-37a5-49b2-b17a-bf95d0d8d23c {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 680.923298] env[63372]: DEBUG nova.network.neutron [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Successfully updated port: fff45461-b9dc-4d60-989d-c7460bc0becb {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 681.006439] env[63372]: DEBUG nova.network.neutron [req-55254c9b-1b75-4b42-8a60-3d6b0e7ae2d7 req-5726b214-3ddd-412d-a2c2-1455233879ca service nova] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Updated VIF entry in instance network info cache for port e018b1c1-a38b-4542-a2b4-bf87f12453ae. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 681.007138] env[63372]: DEBUG nova.network.neutron [req-55254c9b-1b75-4b42-8a60-3d6b0e7ae2d7 req-5726b214-3ddd-412d-a2c2-1455233879ca service nova] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Updating instance_info_cache with network_info: [{"id": "e018b1c1-a38b-4542-a2b4-bf87f12453ae", "address": "fa:16:3e:cf:8c:50", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.85", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape018b1c1-a3", "ovs_interfaceid": "e018b1c1-a38b-4542-a2b4-bf87f12453ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.023558] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023767, 'name': CreateVM_Task, 'duration_secs': 0.989386} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.023558] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 681.024252] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.024417] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.024728] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 681.025013] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6de8356c-9cd1-4664-a5dc-10b778cb4bed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.029614] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 681.029614] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52ba2ebe-a1ae-041a-7897-b498a561cebf" [ 681.029614] env[63372]: _type = "Task" [ 681.029614] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.037075] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ba2ebe-a1ae-041a-7897-b498a561cebf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.166263] env[63372]: DEBUG nova.compute.utils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 681.167640] env[63372]: DEBUG nova.compute.manager [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 681.167853] env[63372]: DEBUG nova.network.neutron [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 681.211936] env[63372]: DEBUG nova.policy [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0cc3768e4e8c47a88565a21b7f05ad02', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '98581685387a4f1499ae6ed378af982c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 681.428304] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "refresh_cache-c4718797-aa86-4ec0-94d3-6480bd6aa898" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.428425] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "refresh_cache-c4718797-aa86-4ec0-94d3-6480bd6aa898" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.428590] env[63372]: DEBUG nova.network.neutron [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 681.477948] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a1acfaf-679a-4e6e-8e99-d547cb6f9e59 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.491020] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e253d034-dbe9-44a6-afb4-a698b719a23d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.522032] env[63372]: DEBUG oslo_concurrency.lockutils [req-55254c9b-1b75-4b42-8a60-3d6b0e7ae2d7 req-5726b214-3ddd-412d-a2c2-1455233879ca service nova] Releasing lock "refresh_cache-5819c38e-2cf9-4d16-b28a-5f23d35c3d44" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.523571] env[63372]: DEBUG nova.network.neutron [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Successfully created port: e1486286-9147-4e5e-88c1-60893a3338dd {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 681.526594] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b2af1a-811a-46a7-a8fb-e4646fcc9960 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.536437] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db65a6ff-8f3d-42f3-8590-384d9b022b74 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.543776] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ba2ebe-a1ae-041a-7897-b498a561cebf, 'name': SearchDatastore_Task, 'duration_secs': 0.06033} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.544111] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.544469] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 681.544637] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.544780] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.544960] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 681.545199] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21d0dca5-1da6-4954-9ca8-fba9adacf93d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.558854] env[63372]: DEBUG nova.compute.provider_tree [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 681.566340] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 681.567453] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 681.567453] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09ccb8d1-6c1a-43a8-befe-e5bab621cb48 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.573177] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 681.573177] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5290c41d-7d8c-2811-7e04-0ccdfe86be21" [ 681.573177] env[63372]: _type = "Task" [ 681.573177] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.581391] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5290c41d-7d8c-2811-7e04-0ccdfe86be21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.670850] env[63372]: DEBUG nova.compute.manager [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 681.965251] env[63372]: DEBUG nova.network.neutron [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.088588] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5290c41d-7d8c-2811-7e04-0ccdfe86be21, 'name': SearchDatastore_Task, 'duration_secs': 0.008062} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.090187] env[63372]: DEBUG nova.scheduler.client.report [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 60 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 682.090426] env[63372]: DEBUG nova.compute.provider_tree [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 60 to 61 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 682.090603] env[63372]: DEBUG nova.compute.provider_tree [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 682.093919] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84e7f56b-b392-42d2-a411-6a508fd89863 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.099779] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 682.099779] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5224d85f-5c1a-728a-a794-c3a3af563305" [ 682.099779] env[63372]: _type = "Task" [ 682.099779] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.107889] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5224d85f-5c1a-728a-a794-c3a3af563305, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.110458] env[63372]: DEBUG nova.network.neutron [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Updating instance_info_cache with network_info: [{"id": "fff45461-b9dc-4d60-989d-c7460bc0becb", "address": "fa:16:3e:77:91:d7", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfff45461-b9", "ovs_interfaceid": "fff45461-b9dc-4d60-989d-c7460bc0becb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.255066] env[63372]: DEBUG nova.compute.manager [req-a56ea950-cd26-42b1-aa77-3cfaac8c1c08 req-584b74d9-b258-45bc-933c-d62a7f995094 service nova] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Received event network-vif-plugged-fff45461-b9dc-4d60-989d-c7460bc0becb {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 682.255363] env[63372]: DEBUG oslo_concurrency.lockutils [req-a56ea950-cd26-42b1-aa77-3cfaac8c1c08 req-584b74d9-b258-45bc-933c-d62a7f995094 service nova] Acquiring lock "c4718797-aa86-4ec0-94d3-6480bd6aa898-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.255632] env[63372]: DEBUG oslo_concurrency.lockutils [req-a56ea950-cd26-42b1-aa77-3cfaac8c1c08 req-584b74d9-b258-45bc-933c-d62a7f995094 service nova] Lock "c4718797-aa86-4ec0-94d3-6480bd6aa898-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.255914] env[63372]: DEBUG oslo_concurrency.lockutils [req-a56ea950-cd26-42b1-aa77-3cfaac8c1c08 req-584b74d9-b258-45bc-933c-d62a7f995094 service nova] Lock "c4718797-aa86-4ec0-94d3-6480bd6aa898-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.256169] env[63372]: DEBUG nova.compute.manager [req-a56ea950-cd26-42b1-aa77-3cfaac8c1c08 req-584b74d9-b258-45bc-933c-d62a7f995094 service nova] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] No waiting events found dispatching network-vif-plugged-fff45461-b9dc-4d60-989d-c7460bc0becb {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 682.256415] env[63372]: WARNING nova.compute.manager [req-a56ea950-cd26-42b1-aa77-3cfaac8c1c08 req-584b74d9-b258-45bc-933c-d62a7f995094 service nova] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Received unexpected event network-vif-plugged-fff45461-b9dc-4d60-989d-c7460bc0becb for instance with vm_state building and task_state spawning. [ 682.256643] env[63372]: DEBUG nova.compute.manager [req-a56ea950-cd26-42b1-aa77-3cfaac8c1c08 req-584b74d9-b258-45bc-933c-d62a7f995094 service nova] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Received event network-changed-fff45461-b9dc-4d60-989d-c7460bc0becb {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 682.256963] env[63372]: DEBUG nova.compute.manager [req-a56ea950-cd26-42b1-aa77-3cfaac8c1c08 req-584b74d9-b258-45bc-933c-d62a7f995094 service nova] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Refreshing instance network info cache due to event network-changed-fff45461-b9dc-4d60-989d-c7460bc0becb. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 682.257222] env[63372]: DEBUG oslo_concurrency.lockutils [req-a56ea950-cd26-42b1-aa77-3cfaac8c1c08 req-584b74d9-b258-45bc-933c-d62a7f995094 service nova] Acquiring lock "refresh_cache-c4718797-aa86-4ec0-94d3-6480bd6aa898" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.599493] env[63372]: DEBUG oslo_concurrency.lockutils [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.937s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.600456] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.213s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.601907] env[63372]: INFO nova.compute.claims [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.614060] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "refresh_cache-c4718797-aa86-4ec0-94d3-6480bd6aa898" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.614330] env[63372]: DEBUG nova.compute.manager [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Instance network_info: |[{"id": "fff45461-b9dc-4d60-989d-c7460bc0becb", "address": "fa:16:3e:77:91:d7", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfff45461-b9", "ovs_interfaceid": "fff45461-b9dc-4d60-989d-c7460bc0becb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 682.614547] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5224d85f-5c1a-728a-a794-c3a3af563305, 'name': SearchDatastore_Task, 'duration_secs': 0.009165} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.614707] env[63372]: DEBUG oslo_concurrency.lockutils [req-a56ea950-cd26-42b1-aa77-3cfaac8c1c08 req-584b74d9-b258-45bc-933c-d62a7f995094 service nova] Acquired lock "refresh_cache-c4718797-aa86-4ec0-94d3-6480bd6aa898" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.614880] env[63372]: DEBUG nova.network.neutron [req-a56ea950-cd26-42b1-aa77-3cfaac8c1c08 req-584b74d9-b258-45bc-933c-d62a7f995094 service nova] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Refreshing network info cache for port fff45461-b9dc-4d60-989d-c7460bc0becb {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 682.616731] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:91:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60badc2d-69d2-467d-a92e-98511f5cb0b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fff45461-b9dc-4d60-989d-c7460bc0becb', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 682.623422] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Creating folder: Project (2d0c0f315749429dbae22a72ae82b1b2). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 682.623700] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.623948] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 5819c38e-2cf9-4d16-b28a-5f23d35c3d44/5819c38e-2cf9-4d16-b28a-5f23d35c3d44.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 682.625453] env[63372]: INFO nova.scheduler.client.report [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Deleted allocations for instance 779e3559-37a5-49b2-b17a-bf95d0d8d23c [ 682.628595] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fcf34646-feb6-4b23-9dec-1aaa2cdeb045 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.630185] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b252b1e5-5fe7-4074-895e-b598df4fabe0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.641019] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 682.641019] env[63372]: value = "task-1023770" [ 682.641019] env[63372]: _type = "Task" [ 682.641019] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.646064] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Created folder: Project (2d0c0f315749429dbae22a72ae82b1b2) in parent group-v227230. [ 682.646295] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Creating folder: Instances. Parent ref: group-v227270. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 682.646885] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-70ba7dde-b11c-42b6-bd59-917000805cba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.652368] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023770, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.659808] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Created folder: Instances in parent group-v227270. [ 682.660057] env[63372]: DEBUG oslo.service.loopingcall [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 682.660898] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 682.660898] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4e373663-745b-4880-a2c8-7f4e9a0aa45a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.682031] env[63372]: DEBUG nova.compute.manager [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 682.688030] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 682.688030] env[63372]: value = "task-1023772" [ 682.688030] env[63372]: _type = "Task" [ 682.688030] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.698982] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023772, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.711023] env[63372]: DEBUG nova.virt.hardware [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 682.711277] env[63372]: DEBUG nova.virt.hardware [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 682.711442] env[63372]: DEBUG nova.virt.hardware [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 682.711614] env[63372]: DEBUG nova.virt.hardware [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 682.711758] env[63372]: DEBUG nova.virt.hardware [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 682.711902] env[63372]: DEBUG nova.virt.hardware [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 682.712118] env[63372]: DEBUG nova.virt.hardware [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 682.712278] env[63372]: DEBUG nova.virt.hardware [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 682.712859] env[63372]: DEBUG nova.virt.hardware [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 682.712859] env[63372]: DEBUG nova.virt.hardware [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 682.712859] env[63372]: DEBUG nova.virt.hardware [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 682.717018] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8476543f-feeb-4b69-b0a3-c761f6e2a019 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.724463] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba8e2dd2-1efd-4a84-a41d-a9730d044a94 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.977300] env[63372]: DEBUG nova.network.neutron [req-a56ea950-cd26-42b1-aa77-3cfaac8c1c08 req-584b74d9-b258-45bc-933c-d62a7f995094 service nova] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Updated VIF entry in instance network info cache for port fff45461-b9dc-4d60-989d-c7460bc0becb. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 682.977716] env[63372]: DEBUG nova.network.neutron [req-a56ea950-cd26-42b1-aa77-3cfaac8c1c08 req-584b74d9-b258-45bc-933c-d62a7f995094 service nova] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Updating instance_info_cache with network_info: [{"id": "fff45461-b9dc-4d60-989d-c7460bc0becb", "address": "fa:16:3e:77:91:d7", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfff45461-b9", "ovs_interfaceid": "fff45461-b9dc-4d60-989d-c7460bc0becb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.140019] env[63372]: DEBUG oslo_concurrency.lockutils [None req-87702656-08dc-4160-b43a-fc15ef8040b9 tempest-ServerAddressesNegativeTestJSON-336256750 tempest-ServerAddressesNegativeTestJSON-336256750-project-member] Lock "779e3559-37a5-49b2-b17a-bf95d0d8d23c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.717s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.156483] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023770, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472323} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.156882] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 5819c38e-2cf9-4d16-b28a-5f23d35c3d44/5819c38e-2cf9-4d16-b28a-5f23d35c3d44.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 683.157218] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 683.157906] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-334e2fa8-e9ba-407f-8975-4e06651a5c9a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.167240] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 683.167240] env[63372]: value = "task-1023773" [ 683.167240] env[63372]: _type = "Task" [ 683.167240] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.180261] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023773, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.199251] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023772, 'name': CreateVM_Task, 'duration_secs': 0.353862} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.199429] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 683.200147] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.200309] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.200642] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 683.201713] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-547bf3b0-3cf0-4ff8-996e-355f3fec307c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.206653] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 683.206653] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]526ff10e-b999-caa9-9498-e90960bc6e00" [ 683.206653] env[63372]: _type = "Task" [ 683.206653] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.213887] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526ff10e-b999-caa9-9498-e90960bc6e00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.321464] env[63372]: DEBUG nova.network.neutron [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Successfully updated port: e1486286-9147-4e5e-88c1-60893a3338dd {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 683.480913] env[63372]: DEBUG oslo_concurrency.lockutils [req-a56ea950-cd26-42b1-aa77-3cfaac8c1c08 req-584b74d9-b258-45bc-933c-d62a7f995094 service nova] Releasing lock "refresh_cache-c4718797-aa86-4ec0-94d3-6480bd6aa898" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.676645] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023773, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073048} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.679223] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 683.680368] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ca0563-9088-448e-8b91-da7b985d8c96 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.704164] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] 5819c38e-2cf9-4d16-b28a-5f23d35c3d44/5819c38e-2cf9-4d16-b28a-5f23d35c3d44.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 683.706930] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c89b385-53f4-43d5-8dae-0aff2092a9da {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.740029] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526ff10e-b999-caa9-9498-e90960bc6e00, 'name': SearchDatastore_Task, 'duration_secs': 0.009242} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.740999] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.741391] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 683.741785] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.742114] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.744094] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 683.744094] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 683.744094] env[63372]: value = "task-1023775" [ 683.744094] env[63372]: _type = "Task" [ 683.744094] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.744094] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b047fa9-aaf3-468d-9214-87c54b584628 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.758713] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023775, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.759205] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 683.759451] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 683.760281] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ecb714a-f2ea-4fd8-8be4-93790db92acb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.765857] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 683.765857] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]522e8023-e861-9c93-5a22-c7f9432a7503" [ 683.765857] env[63372]: _type = "Task" [ 683.765857] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.774932] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522e8023-e861-9c93-5a22-c7f9432a7503, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.824607] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.824607] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.824607] env[63372]: DEBUG nova.network.neutron [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 684.011155] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20fd27de-e512-499e-b0c1-f08063b99046 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.018743] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fdf2e6-87f9-45a7-9f16-8dee36fc2d5f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.050636] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c3d0bf-f35f-4874-998b-8b6c2f12c5fa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.058438] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffd0f08-23e2-457e-8036-841989ed09c7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.071979] env[63372]: DEBUG nova.compute.provider_tree [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.257718] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023775, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.277194] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522e8023-e861-9c93-5a22-c7f9432a7503, 'name': SearchDatastore_Task, 'duration_secs': 0.008723} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.280773] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3aa7b1d8-5c1c-4fac-9e92-3524f4c951d1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.283887] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 684.283887] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5265f707-81d0-2b8f-f389-1c57efac02a5" [ 684.283887] env[63372]: _type = "Task" [ 684.283887] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.293321] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5265f707-81d0-2b8f-f389-1c57efac02a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.382666] env[63372]: DEBUG nova.network.neutron [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.425158] env[63372]: DEBUG nova.compute.manager [req-bbe526d3-ed5d-4130-8b27-659a2fa9d7b5 req-a244d729-cfcf-42d7-827a-5e1403cdd6a5 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Received event network-vif-plugged-e1486286-9147-4e5e-88c1-60893a3338dd {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 684.425158] env[63372]: DEBUG oslo_concurrency.lockutils [req-bbe526d3-ed5d-4130-8b27-659a2fa9d7b5 req-a244d729-cfcf-42d7-827a-5e1403cdd6a5 service nova] Acquiring lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.425158] env[63372]: DEBUG oslo_concurrency.lockutils [req-bbe526d3-ed5d-4130-8b27-659a2fa9d7b5 req-a244d729-cfcf-42d7-827a-5e1403cdd6a5 service nova] Lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.425610] env[63372]: DEBUG oslo_concurrency.lockutils [req-bbe526d3-ed5d-4130-8b27-659a2fa9d7b5 req-a244d729-cfcf-42d7-827a-5e1403cdd6a5 service nova] Lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.428010] env[63372]: DEBUG nova.compute.manager [req-bbe526d3-ed5d-4130-8b27-659a2fa9d7b5 req-a244d729-cfcf-42d7-827a-5e1403cdd6a5 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] No waiting events found dispatching network-vif-plugged-e1486286-9147-4e5e-88c1-60893a3338dd {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 684.428010] env[63372]: WARNING nova.compute.manager [req-bbe526d3-ed5d-4130-8b27-659a2fa9d7b5 req-a244d729-cfcf-42d7-827a-5e1403cdd6a5 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Received unexpected event network-vif-plugged-e1486286-9147-4e5e-88c1-60893a3338dd for instance with vm_state building and task_state spawning. [ 684.428010] env[63372]: DEBUG nova.compute.manager [req-bbe526d3-ed5d-4130-8b27-659a2fa9d7b5 req-a244d729-cfcf-42d7-827a-5e1403cdd6a5 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Received event network-changed-e1486286-9147-4e5e-88c1-60893a3338dd {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 684.428010] env[63372]: DEBUG nova.compute.manager [req-bbe526d3-ed5d-4130-8b27-659a2fa9d7b5 req-a244d729-cfcf-42d7-827a-5e1403cdd6a5 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Refreshing instance network info cache due to event network-changed-e1486286-9147-4e5e-88c1-60893a3338dd. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 684.428010] env[63372]: DEBUG oslo_concurrency.lockutils [req-bbe526d3-ed5d-4130-8b27-659a2fa9d7b5 req-a244d729-cfcf-42d7-827a-5e1403cdd6a5 service nova] Acquiring lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.580783] env[63372]: DEBUG nova.scheduler.client.report [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.686318] env[63372]: DEBUG nova.network.neutron [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance_info_cache with network_info: [{"id": "e1486286-9147-4e5e-88c1-60893a3338dd", "address": "fa:16:3e:6c:92:ac", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1486286-91", "ovs_interfaceid": "e1486286-9147-4e5e-88c1-60893a3338dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.762737] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023775, 'name': ReconfigVM_Task, 'duration_secs': 0.863423} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.763136] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Reconfigured VM instance instance-00000022 to attach disk [datastore2] 5819c38e-2cf9-4d16-b28a-5f23d35c3d44/5819c38e-2cf9-4d16-b28a-5f23d35c3d44.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 684.763694] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa3a2b69-cc4b-41af-a32b-62adb47f876d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.771320] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 684.771320] env[63372]: value = "task-1023776" [ 684.771320] env[63372]: _type = "Task" [ 684.771320] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.781170] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023776, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.794157] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5265f707-81d0-2b8f-f389-1c57efac02a5, 'name': SearchDatastore_Task, 'duration_secs': 0.009931} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.794471] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.794846] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] c4718797-aa86-4ec0-94d3-6480bd6aa898/c4718797-aa86-4ec0-94d3-6480bd6aa898.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 684.795177] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-be3cd2c3-de8a-48b9-80e5-b9e33a65aafb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.802500] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 684.802500] env[63372]: value = "task-1023777" [ 684.802500] env[63372]: _type = "Task" [ 684.802500] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.813155] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023777, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.085676] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.086425] env[63372]: DEBUG nova.compute.manager [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 685.090512] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.164s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.092714] env[63372]: INFO nova.compute.claims [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 685.189987] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.191031] env[63372]: DEBUG nova.compute.manager [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Instance network_info: |[{"id": "e1486286-9147-4e5e-88c1-60893a3338dd", "address": "fa:16:3e:6c:92:ac", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1486286-91", "ovs_interfaceid": "e1486286-9147-4e5e-88c1-60893a3338dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 685.191288] env[63372]: DEBUG oslo_concurrency.lockutils [req-bbe526d3-ed5d-4130-8b27-659a2fa9d7b5 req-a244d729-cfcf-42d7-827a-5e1403cdd6a5 service nova] Acquired lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.191866] env[63372]: DEBUG nova.network.neutron [req-bbe526d3-ed5d-4130-8b27-659a2fa9d7b5 req-a244d729-cfcf-42d7-827a-5e1403cdd6a5 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Refreshing network info cache for port e1486286-9147-4e5e-88c1-60893a3338dd {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 685.193242] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:92:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e1486286-9147-4e5e-88c1-60893a3338dd', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 685.202435] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Creating folder: Project (98581685387a4f1499ae6ed378af982c). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 685.203260] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8be7fb8e-205d-40fa-96d0-95d351c7941e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.218475] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Created folder: Project (98581685387a4f1499ae6ed378af982c) in parent group-v227230. [ 685.218684] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Creating folder: Instances. Parent ref: group-v227273. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 685.219430] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b9e535a8-e528-40bc-b5fa-3f80574ceced {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.233122] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Created folder: Instances in parent group-v227273. [ 685.233589] env[63372]: DEBUG oslo.service.loopingcall [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 685.234063] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 685.234063] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-51891f10-0d7d-422a-9b73-74038e2a974f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.255314] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 685.255314] env[63372]: value = "task-1023780" [ 685.255314] env[63372]: _type = "Task" [ 685.255314] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.267045] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023780, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.281436] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023776, 'name': Rename_Task, 'duration_secs': 0.173266} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.281729] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 685.281966] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cde4fe91-2068-4a52-a349-992fed2bb31c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.291512] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 685.291512] env[63372]: value = "task-1023781" [ 685.291512] env[63372]: _type = "Task" [ 685.291512] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.303030] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023781, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.316740] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023777, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.598438] env[63372]: DEBUG nova.compute.utils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 685.603827] env[63372]: DEBUG nova.compute.manager [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 685.604133] env[63372]: DEBUG nova.network.neutron [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 685.659868] env[63372]: DEBUG nova.policy [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '59b6b66e2d2d44689b5000b7b5708b76', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1c06937edfd41a1851d424d6603cec8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 685.768199] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023780, 'name': CreateVM_Task, 'duration_secs': 0.479768} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.768468] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 685.769181] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.769350] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.769676] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 685.770081] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fddb453-7ca9-4bd3-8519-99b72d46e63c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.780636] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 685.780636] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525aa556-b15f-1a93-ffbe-6c25363610e8" [ 685.780636] env[63372]: _type = "Task" [ 685.780636] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.785613] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525aa556-b15f-1a93-ffbe-6c25363610e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.801246] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023781, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.813449] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023777, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.702038} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.813598] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] c4718797-aa86-4ec0-94d3-6480bd6aa898/c4718797-aa86-4ec0-94d3-6480bd6aa898.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 685.813801] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 685.814084] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c431a3df-e340-4eb1-8dcb-7f6715b3a104 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.823213] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 685.823213] env[63372]: value = "task-1023783" [ 685.823213] env[63372]: _type = "Task" [ 685.823213] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.839194] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023783, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.068581] env[63372]: DEBUG nova.network.neutron [req-bbe526d3-ed5d-4130-8b27-659a2fa9d7b5 req-a244d729-cfcf-42d7-827a-5e1403cdd6a5 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updated VIF entry in instance network info cache for port e1486286-9147-4e5e-88c1-60893a3338dd. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 686.069055] env[63372]: DEBUG nova.network.neutron [req-bbe526d3-ed5d-4130-8b27-659a2fa9d7b5 req-a244d729-cfcf-42d7-827a-5e1403cdd6a5 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance_info_cache with network_info: [{"id": "e1486286-9147-4e5e-88c1-60893a3338dd", "address": "fa:16:3e:6c:92:ac", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1486286-91", "ovs_interfaceid": "e1486286-9147-4e5e-88c1-60893a3338dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.106194] env[63372]: DEBUG nova.compute.manager [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 686.118468] env[63372]: DEBUG nova.network.neutron [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Successfully created port: 7cb73607-f39d-49bc-ac64-cec7fd400af3 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.296782] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525aa556-b15f-1a93-ffbe-6c25363610e8, 'name': SearchDatastore_Task, 'duration_secs': 0.020415} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.306376] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.306673] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 686.306961] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.307183] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.307376] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 686.308556] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1903dff1-8d9f-4c1e-9b5c-dd10793a64a8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.319339] env[63372]: DEBUG oslo_vmware.api [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023781, 'name': PowerOnVM_Task, 'duration_secs': 0.812978} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.320768] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 686.320768] env[63372]: INFO nova.compute.manager [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Took 8.69 seconds to spawn the instance on the hypervisor. [ 686.320768] env[63372]: DEBUG nova.compute.manager [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 686.321030] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-258e956f-e0cb-4237-a42f-13545ee5e272 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.327618] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 686.327798] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 686.330109] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38f3b40a-9129-4eaa-87b5-06bab603813c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.344131] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 686.344131] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5224884e-0afd-5ffa-1787-c6a60a455a9e" [ 686.344131] env[63372]: _type = "Task" [ 686.344131] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.351269] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023783, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.210273} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.360433] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 686.361954] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e366903-f264-4923-af1e-8c37aa56dc3f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.374276] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5224884e-0afd-5ffa-1787-c6a60a455a9e, 'name': SearchDatastore_Task, 'duration_secs': 0.022383} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.395600] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Reconfiguring VM instance instance-00000023 to attach disk [datastore2] c4718797-aa86-4ec0-94d3-6480bd6aa898/c4718797-aa86-4ec0-94d3-6480bd6aa898.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 686.399284] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1ed625a-4195-436d-ab98-6f86507d3490 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.401665] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8659b7e1-33c3-435e-a6d3-7320c369499b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.420868] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 686.420868] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52d34591-3da0-f40c-55ef-ca2f6c28f840" [ 686.420868] env[63372]: _type = "Task" [ 686.420868] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.425830] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 686.425830] env[63372]: value = "task-1023784" [ 686.425830] env[63372]: _type = "Task" [ 686.425830] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.436483] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52d34591-3da0-f40c-55ef-ca2f6c28f840, 'name': SearchDatastore_Task, 'duration_secs': 0.010092} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.437164] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.437472] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5/19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 686.437756] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1f606406-3ec9-42a7-bac1-d56867e0560b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.446046] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023784, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.453020] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 686.453020] env[63372]: value = "task-1023785" [ 686.453020] env[63372]: _type = "Task" [ 686.453020] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.460675] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1023785, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.571903] env[63372]: DEBUG oslo_concurrency.lockutils [req-bbe526d3-ed5d-4130-8b27-659a2fa9d7b5 req-a244d729-cfcf-42d7-827a-5e1403cdd6a5 service nova] Releasing lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.574207] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aebb13c-4361-48c7-8005-35ac6942193a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.583098] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2648612-4ef9-46ef-8433-a2f1315c1d2b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.617901] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05458d2-22f8-4b8d-92bc-0799412b45c4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.629714] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca92423-68c5-4eeb-8eae-9fbb3f9cd88e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.646517] env[63372]: DEBUG nova.compute.provider_tree [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.856213] env[63372]: INFO nova.compute.manager [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Took 37.97 seconds to build instance. [ 686.941291] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023784, 'name': ReconfigVM_Task, 'duration_secs': 0.330952} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.941634] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Reconfigured VM instance instance-00000023 to attach disk [datastore2] c4718797-aa86-4ec0-94d3-6480bd6aa898/c4718797-aa86-4ec0-94d3-6480bd6aa898.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 686.942563] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7036801b-96af-45fd-b80a-7999a09153c4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.949765] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 686.949765] env[63372]: value = "task-1023786" [ 686.949765] env[63372]: _type = "Task" [ 686.949765] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.964724] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023786, 'name': Rename_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.964868] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1023785, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.136832] env[63372]: DEBUG nova.compute.manager [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 687.150123] env[63372]: DEBUG nova.scheduler.client.report [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.164264] env[63372]: DEBUG nova.virt.hardware [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 687.164837] env[63372]: DEBUG nova.virt.hardware [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 687.164837] env[63372]: DEBUG nova.virt.hardware [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.164934] env[63372]: DEBUG nova.virt.hardware [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 687.165034] env[63372]: DEBUG nova.virt.hardware [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.165209] env[63372]: DEBUG nova.virt.hardware [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 687.165420] env[63372]: DEBUG nova.virt.hardware [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 687.165595] env[63372]: DEBUG nova.virt.hardware [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 687.165762] env[63372]: DEBUG nova.virt.hardware [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 687.165937] env[63372]: DEBUG nova.virt.hardware [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 687.166161] env[63372]: DEBUG nova.virt.hardware [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 687.167114] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd274e1-dc1e-4536-90a7-0db2a14fb4f8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.177064] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2038d64-f882-4d77-b11c-80865663135f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.362615] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca96d7e3-e21d-4fef-981a-235a25e42430 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "5819c38e-2cf9-4d16-b28a-5f23d35c3d44" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.102s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.466017] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023786, 'name': Rename_Task, 'duration_secs': 0.278789} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.469278] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 687.469581] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1023785, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.469795] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f8cea27-865e-4029-828b-23ea4199c2f3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.476217] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 687.476217] env[63372]: value = "task-1023787" [ 687.476217] env[63372]: _type = "Task" [ 687.476217] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.484532] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023787, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.629553] env[63372]: DEBUG nova.compute.manager [req-c0e61cdb-24d5-4a68-974b-dca291fc9bb7 req-da84b12c-84df-4f8a-8a43-24e1d4435091 service nova] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Received event network-vif-plugged-7cb73607-f39d-49bc-ac64-cec7fd400af3 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 687.629773] env[63372]: DEBUG oslo_concurrency.lockutils [req-c0e61cdb-24d5-4a68-974b-dca291fc9bb7 req-da84b12c-84df-4f8a-8a43-24e1d4435091 service nova] Acquiring lock "276c3490-240a-499b-ba64-4b02510b169c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.629978] env[63372]: DEBUG oslo_concurrency.lockutils [req-c0e61cdb-24d5-4a68-974b-dca291fc9bb7 req-da84b12c-84df-4f8a-8a43-24e1d4435091 service nova] Lock "276c3490-240a-499b-ba64-4b02510b169c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.630164] env[63372]: DEBUG oslo_concurrency.lockutils [req-c0e61cdb-24d5-4a68-974b-dca291fc9bb7 req-da84b12c-84df-4f8a-8a43-24e1d4435091 service nova] Lock "276c3490-240a-499b-ba64-4b02510b169c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.630392] env[63372]: DEBUG nova.compute.manager [req-c0e61cdb-24d5-4a68-974b-dca291fc9bb7 req-da84b12c-84df-4f8a-8a43-24e1d4435091 service nova] [instance: 276c3490-240a-499b-ba64-4b02510b169c] No waiting events found dispatching network-vif-plugged-7cb73607-f39d-49bc-ac64-cec7fd400af3 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 687.630486] env[63372]: WARNING nova.compute.manager [req-c0e61cdb-24d5-4a68-974b-dca291fc9bb7 req-da84b12c-84df-4f8a-8a43-24e1d4435091 service nova] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Received unexpected event network-vif-plugged-7cb73607-f39d-49bc-ac64-cec7fd400af3 for instance with vm_state building and task_state spawning. [ 687.656032] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.566s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.656584] env[63372]: DEBUG nova.compute.manager [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 687.659329] env[63372]: DEBUG oslo_concurrency.lockutils [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.612s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.661426] env[63372]: INFO nova.compute.claims [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.723592] env[63372]: DEBUG nova.network.neutron [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Successfully updated port: 7cb73607-f39d-49bc-ac64-cec7fd400af3 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 687.865862] env[63372]: DEBUG nova.compute.manager [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 687.965440] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1023785, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.495821} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.965712] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5/19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 687.965985] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 687.966288] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7f7c066-418b-408a-a2b7-64f0a3b16e80 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.974552] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 687.974552] env[63372]: value = "task-1023788" [ 687.974552] env[63372]: _type = "Task" [ 687.974552] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.986924] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023787, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.989831] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1023788, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.018092] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "7cf6ac9b-4703-4143-a22c-abc5528ce5ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.018418] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "7cf6ac9b-4703-4143-a22c-abc5528ce5ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.165236] env[63372]: DEBUG nova.compute.utils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 688.168798] env[63372]: DEBUG nova.compute.manager [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 688.168798] env[63372]: DEBUG nova.network.neutron [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 688.219708] env[63372]: DEBUG nova.policy [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7bb41d57f8d4da582818b2f28a86fc9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2319173a3c445ca93ae2510141be32c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 688.226053] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Acquiring lock "refresh_cache-276c3490-240a-499b-ba64-4b02510b169c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.226262] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Acquired lock "refresh_cache-276c3490-240a-499b-ba64-4b02510b169c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.226435] env[63372]: DEBUG nova.network.neutron [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 688.386904] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.489978] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1023788, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074822} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.491092] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 688.491431] env[63372]: DEBUG oslo_vmware.api [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023787, 'name': PowerOnVM_Task, 'duration_secs': 0.574851} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.492152] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142efc6e-7356-4ba9-974f-9d8f772b2485 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.495877] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 688.495877] env[63372]: INFO nova.compute.manager [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Took 8.33 seconds to spawn the instance on the hypervisor. [ 688.495877] env[63372]: DEBUG nova.compute.manager [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 688.496592] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694c9e61-1a4e-403d-b103-b2b18c964595 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.516660] env[63372]: DEBUG nova.network.neutron [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Successfully created port: c8f78ebe-604d-4414-a16d-ae25256f2730 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 688.529311] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5/19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 688.529311] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-495cf82f-c03a-4ec4-b141-70c0dcee10fd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.549283] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 688.549283] env[63372]: value = "task-1023789" [ 688.549283] env[63372]: _type = "Task" [ 688.549283] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.560932] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1023789, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.675122] env[63372]: DEBUG nova.compute.manager [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 688.781530] env[63372]: DEBUG nova.network.neutron [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.983300] env[63372]: DEBUG nova.network.neutron [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Updating instance_info_cache with network_info: [{"id": "7cb73607-f39d-49bc-ac64-cec7fd400af3", "address": "fa:16:3e:f3:99:e2", "network": {"id": "1dbfbebb-8025-4b09-9740-5562dccd7ef2", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1579060607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1c06937edfd41a1851d424d6603cec8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae70d41-6ebf-472a-8504-6530eb37ea41", "external-id": "nsx-vlan-transportzone-576", "segmentation_id": 576, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cb73607-f3", "ovs_interfaceid": "7cb73607-f39d-49bc-ac64-cec7fd400af3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.038416] env[63372]: INFO nova.compute.manager [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Took 38.24 seconds to build instance. [ 689.059136] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c40ecd55-1ae2-4908-8e2d-7ed9a71047b1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.064595] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1023789, 'name': ReconfigVM_Task, 'duration_secs': 0.35176} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.065195] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Reconfigured VM instance instance-00000024 to attach disk [datastore2] 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5/19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 689.065787] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-06ca23e2-82b3-45c6-8452-c8843e352009 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.070085] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8377618f-b7f2-43c0-a824-1088be8ee797 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.074874] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 689.074874] env[63372]: value = "task-1023790" [ 689.074874] env[63372]: _type = "Task" [ 689.074874] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.103487] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc2d021-2153-4b61-b1de-48e3f1fb7edc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.109543] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1023790, 'name': Rename_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.117108] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81f9158-290f-4044-b280-1bd50158e2bd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.131562] env[63372]: DEBUG nova.compute.provider_tree [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 689.486147] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Releasing lock "refresh_cache-276c3490-240a-499b-ba64-4b02510b169c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.486444] env[63372]: DEBUG nova.compute.manager [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Instance network_info: |[{"id": "7cb73607-f39d-49bc-ac64-cec7fd400af3", "address": "fa:16:3e:f3:99:e2", "network": {"id": "1dbfbebb-8025-4b09-9740-5562dccd7ef2", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1579060607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1c06937edfd41a1851d424d6603cec8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae70d41-6ebf-472a-8504-6530eb37ea41", "external-id": "nsx-vlan-transportzone-576", "segmentation_id": 576, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cb73607-f3", "ovs_interfaceid": "7cb73607-f39d-49bc-ac64-cec7fd400af3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 689.486843] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:99:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cae70d41-6ebf-472a-8504-6530eb37ea41', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7cb73607-f39d-49bc-ac64-cec7fd400af3', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 689.494326] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Creating folder: Project (b1c06937edfd41a1851d424d6603cec8). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 689.494586] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f08af4b6-2fdd-4bba-baad-980328deccb4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.505776] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Created folder: Project (b1c06937edfd41a1851d424d6603cec8) in parent group-v227230. [ 689.505952] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Creating folder: Instances. Parent ref: group-v227277. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 689.506239] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53953d8b-9c02-4503-b79d-7ec0a86c0c04 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.516455] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Created folder: Instances in parent group-v227277. [ 689.516707] env[63372]: DEBUG oslo.service.loopingcall [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.516920] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 689.517103] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-54103f8b-2abe-4f98-b717-8a6ce91bb9ff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.535571] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 689.535571] env[63372]: value = "task-1023793" [ 689.535571] env[63372]: _type = "Task" [ 689.535571] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.540389] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6aa8381-fab1-4209-997f-0b59662a6043 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "c4718797-aa86-4ec0-94d3-6480bd6aa898" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.967s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.544346] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023793, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.587188] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1023790, 'name': Rename_Task, 'duration_secs': 0.233082} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.587493] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 689.587701] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea6b121b-4d0e-4e00-a570-9b49232eccbf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.594786] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 689.594786] env[63372]: value = "task-1023794" [ 689.594786] env[63372]: _type = "Task" [ 689.594786] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.607296] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1023794, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.653692] env[63372]: ERROR nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [req-334ea35f-ffec-4d72-b665-70379eebfcc3] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-334ea35f-ffec-4d72-b665-70379eebfcc3"}]} [ 689.671367] env[63372]: DEBUG nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Refreshing inventories for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 689.684615] env[63372]: DEBUG nova.compute.manager [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 689.687628] env[63372]: DEBUG nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Updating ProviderTree inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 689.687921] env[63372]: DEBUG nova.compute.provider_tree [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 689.700994] env[63372]: DEBUG nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Refreshing aggregate associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, aggregates: None {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 689.715319] env[63372]: DEBUG nova.virt.hardware [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 689.715709] env[63372]: DEBUG nova.virt.hardware [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 689.716012] env[63372]: DEBUG nova.virt.hardware [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 689.716462] env[63372]: DEBUG nova.virt.hardware [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 689.717048] env[63372]: DEBUG nova.virt.hardware [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 689.717048] env[63372]: DEBUG nova.virt.hardware [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 689.717320] env[63372]: DEBUG nova.virt.hardware [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 689.717493] env[63372]: DEBUG nova.virt.hardware [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 689.717665] env[63372]: DEBUG nova.virt.hardware [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 689.717828] env[63372]: DEBUG nova.virt.hardware [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 689.718000] env[63372]: DEBUG nova.virt.hardware [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 689.719008] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc7bf68-281b-4f2e-8fb5-5701f984e7e9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.723999] env[63372]: DEBUG nova.compute.manager [req-860d6b7d-0b27-409c-9d81-571678be8bee req-0c0b4796-3999-4fc8-87b4-1adf15a8e18a service nova] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Received event network-changed-7cb73607-f39d-49bc-ac64-cec7fd400af3 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 689.724291] env[63372]: DEBUG nova.compute.manager [req-860d6b7d-0b27-409c-9d81-571678be8bee req-0c0b4796-3999-4fc8-87b4-1adf15a8e18a service nova] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Refreshing instance network info cache due to event network-changed-7cb73607-f39d-49bc-ac64-cec7fd400af3. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 689.724557] env[63372]: DEBUG oslo_concurrency.lockutils [req-860d6b7d-0b27-409c-9d81-571678be8bee req-0c0b4796-3999-4fc8-87b4-1adf15a8e18a service nova] Acquiring lock "refresh_cache-276c3490-240a-499b-ba64-4b02510b169c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.724701] env[63372]: DEBUG oslo_concurrency.lockutils [req-860d6b7d-0b27-409c-9d81-571678be8bee req-0c0b4796-3999-4fc8-87b4-1adf15a8e18a service nova] Acquired lock "refresh_cache-276c3490-240a-499b-ba64-4b02510b169c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.724856] env[63372]: DEBUG nova.network.neutron [req-860d6b7d-0b27-409c-9d81-571678be8bee req-0c0b4796-3999-4fc8-87b4-1adf15a8e18a service nova] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Refreshing network info cache for port 7cb73607-f39d-49bc-ac64-cec7fd400af3 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 689.727166] env[63372]: DEBUG nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Refreshing trait associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 689.735181] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399fedad-c491-4845-84b8-b27f24f723df {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.048318] env[63372]: DEBUG nova.compute.manager [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 690.050905] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023793, 'name': CreateVM_Task, 'duration_secs': 0.437617} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.051015] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 690.052108] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.052108] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.052288] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 690.052513] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d0ffbfb-dc87-4818-9535-8a037d0dbd5b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.059663] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Waiting for the task: (returnval){ [ 690.059663] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521f92e7-cf3e-db53-85ea-8726293464eb" [ 690.059663] env[63372]: _type = "Task" [ 690.059663] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.073125] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521f92e7-cf3e-db53-85ea-8726293464eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.105946] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1023794, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.107661] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e682938b-5db0-4ad3-863a-ed02848cc2a0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.115964] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f96d88-9fe8-429e-a855-89c6fd80a3be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.154187] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51cec502-2b15-4e08-adb1-41bcf9577aae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.163593] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27bb46b7-02b1-4a3d-b5ad-41e3f15b5467 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.181178] env[63372]: DEBUG nova.compute.provider_tree [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 690.299974] env[63372]: DEBUG nova.network.neutron [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Successfully updated port: c8f78ebe-604d-4414-a16d-ae25256f2730 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 690.481951] env[63372]: DEBUG nova.network.neutron [req-860d6b7d-0b27-409c-9d81-571678be8bee req-0c0b4796-3999-4fc8-87b4-1adf15a8e18a service nova] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Updated VIF entry in instance network info cache for port 7cb73607-f39d-49bc-ac64-cec7fd400af3. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 690.482313] env[63372]: DEBUG nova.network.neutron [req-860d6b7d-0b27-409c-9d81-571678be8bee req-0c0b4796-3999-4fc8-87b4-1adf15a8e18a service nova] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Updating instance_info_cache with network_info: [{"id": "7cb73607-f39d-49bc-ac64-cec7fd400af3", "address": "fa:16:3e:f3:99:e2", "network": {"id": "1dbfbebb-8025-4b09-9740-5562dccd7ef2", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1579060607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1c06937edfd41a1851d424d6603cec8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae70d41-6ebf-472a-8504-6530eb37ea41", "external-id": "nsx-vlan-transportzone-576", "segmentation_id": 576, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cb73607-f3", "ovs_interfaceid": "7cb73607-f39d-49bc-ac64-cec7fd400af3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.568056] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.572386] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521f92e7-cf3e-db53-85ea-8726293464eb, 'name': SearchDatastore_Task, 'duration_secs': 0.014085} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.572654] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.572868] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 690.573098] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.573243] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.573413] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 690.573661] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa332671-41e4-4ab3-9bfe-05bb479546fa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.583560] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 690.583741] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 690.584557] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b8512e4-dc11-47b5-8334-d0209a994835 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.590902] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Waiting for the task: (returnval){ [ 690.590902] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52191eca-704b-a9a0-6508-caab6a63d974" [ 690.590902] env[63372]: _type = "Task" [ 690.590902] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.598897] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52191eca-704b-a9a0-6508-caab6a63d974, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.606399] env[63372]: DEBUG oslo_vmware.api [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1023794, 'name': PowerOnVM_Task, 'duration_secs': 0.632978} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.606637] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 690.606828] env[63372]: INFO nova.compute.manager [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Took 7.93 seconds to spawn the instance on the hypervisor. [ 690.606998] env[63372]: DEBUG nova.compute.manager [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 690.607734] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034d3e6c-c5ce-46bb-b3a0-1c49a4a3562b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.709236] env[63372]: ERROR nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [req-8fa0695d-411b-45e7-90d6-099c31b84e97] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8fa0695d-411b-45e7-90d6-099c31b84e97"}]} [ 690.770509] env[63372]: DEBUG nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Refreshing inventories for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 690.783931] env[63372]: DEBUG nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Updating ProviderTree inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 690.784735] env[63372]: DEBUG nova.compute.provider_tree [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 690.796325] env[63372]: DEBUG nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Refreshing aggregate associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, aggregates: None {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 690.806120] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquiring lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.806120] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquired lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.806120] env[63372]: DEBUG nova.network.neutron [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 690.819335] env[63372]: DEBUG nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Refreshing trait associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 690.985147] env[63372]: DEBUG oslo_concurrency.lockutils [req-860d6b7d-0b27-409c-9d81-571678be8bee req-0c0b4796-3999-4fc8-87b4-1adf15a8e18a service nova] Releasing lock "refresh_cache-276c3490-240a-499b-ba64-4b02510b169c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.101162] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52191eca-704b-a9a0-6508-caab6a63d974, 'name': SearchDatastore_Task, 'duration_secs': 0.00954} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.101958] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd6556b6-0ba6-48d6-9a8a-29c7489d9eba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.108309] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Waiting for the task: (returnval){ [ 691.108309] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fd01db-b60f-9d89-d7eb-b7bb218db0fb" [ 691.108309] env[63372]: _type = "Task" [ 691.108309] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.121158] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fd01db-b60f-9d89-d7eb-b7bb218db0fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.125697] env[63372]: INFO nova.compute.manager [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Took 28.98 seconds to build instance. [ 691.145730] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-754183b2-7b5e-492f-9254-e4aba82609bc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.153802] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc8d812-8c94-46b8-b728-5b1734aba7cf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.187687] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d3391a-8215-4dcd-a357-38f12656b0b3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.197420] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b19f874c-c814-42c9-a540-1072138798a5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.211174] env[63372]: DEBUG nova.compute.provider_tree [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 691.292578] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "757df632-4af2-4eb3-bc6d-8c0812ba850d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.292924] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "757df632-4af2-4eb3-bc6d-8c0812ba850d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.338828] env[63372]: DEBUG nova.network.neutron [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.485325] env[63372]: DEBUG nova.network.neutron [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Updating instance_info_cache with network_info: [{"id": "c8f78ebe-604d-4414-a16d-ae25256f2730", "address": "fa:16:3e:25:30:2b", "network": {"id": "73cbc616-354c-40de-971d-91fb5f0cbbac", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-817872371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2319173a3c445ca93ae2510141be32c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d", "external-id": "nsx-vlan-transportzone-591", "segmentation_id": 591, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8f78ebe-60", "ovs_interfaceid": "c8f78ebe-604d-4414-a16d-ae25256f2730", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.619050] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fd01db-b60f-9d89-d7eb-b7bb218db0fb, 'name': SearchDatastore_Task, 'duration_secs': 0.00979} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.619259] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.619527] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 276c3490-240a-499b-ba64-4b02510b169c/276c3490-240a-499b-ba64-4b02510b169c.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 691.619860] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aab5d2b5-99b5-44ee-a044-fe335ed0ef28 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.627082] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5b76d303-7d7a-4284-adc8-34c63ca89d08 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.650s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.627390] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Waiting for the task: (returnval){ [ 691.627390] env[63372]: value = "task-1023796" [ 691.627390] env[63372]: _type = "Task" [ 691.627390] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.637774] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023796, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.732945] env[63372]: ERROR nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [req-da2df9ac-70cb-4c0f-b8f0-22f0f98a9882] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-da2df9ac-70cb-4c0f-b8f0-22f0f98a9882"}]} [ 691.750386] env[63372]: DEBUG nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Refreshing inventories for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 691.764960] env[63372]: DEBUG nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Updating ProviderTree inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 691.765290] env[63372]: DEBUG nova.compute.provider_tree [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 691.775039] env[63372]: DEBUG nova.compute.manager [req-f2a26a71-5742-45cf-bf10-89d64496a94e req-25abf694-236b-4585-a023-59baeaedbaed service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Received event network-vif-plugged-c8f78ebe-604d-4414-a16d-ae25256f2730 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 691.775244] env[63372]: DEBUG oslo_concurrency.lockutils [req-f2a26a71-5742-45cf-bf10-89d64496a94e req-25abf694-236b-4585-a023-59baeaedbaed service nova] Acquiring lock "d50919ac-3a0b-46ac-a837-ca3e6ad05173-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.775458] env[63372]: DEBUG oslo_concurrency.lockutils [req-f2a26a71-5742-45cf-bf10-89d64496a94e req-25abf694-236b-4585-a023-59baeaedbaed service nova] Lock "d50919ac-3a0b-46ac-a837-ca3e6ad05173-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.775601] env[63372]: DEBUG oslo_concurrency.lockutils [req-f2a26a71-5742-45cf-bf10-89d64496a94e req-25abf694-236b-4585-a023-59baeaedbaed service nova] Lock "d50919ac-3a0b-46ac-a837-ca3e6ad05173-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.775883] env[63372]: DEBUG nova.compute.manager [req-f2a26a71-5742-45cf-bf10-89d64496a94e req-25abf694-236b-4585-a023-59baeaedbaed service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] No waiting events found dispatching network-vif-plugged-c8f78ebe-604d-4414-a16d-ae25256f2730 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 691.776039] env[63372]: WARNING nova.compute.manager [req-f2a26a71-5742-45cf-bf10-89d64496a94e req-25abf694-236b-4585-a023-59baeaedbaed service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Received unexpected event network-vif-plugged-c8f78ebe-604d-4414-a16d-ae25256f2730 for instance with vm_state building and task_state spawning. [ 691.776467] env[63372]: DEBUG nova.compute.manager [req-f2a26a71-5742-45cf-bf10-89d64496a94e req-25abf694-236b-4585-a023-59baeaedbaed service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Received event network-changed-c8f78ebe-604d-4414-a16d-ae25256f2730 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 691.776699] env[63372]: DEBUG nova.compute.manager [req-f2a26a71-5742-45cf-bf10-89d64496a94e req-25abf694-236b-4585-a023-59baeaedbaed service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Refreshing instance network info cache due to event network-changed-c8f78ebe-604d-4414-a16d-ae25256f2730. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 691.776829] env[63372]: DEBUG oslo_concurrency.lockutils [req-f2a26a71-5742-45cf-bf10-89d64496a94e req-25abf694-236b-4585-a023-59baeaedbaed service nova] Acquiring lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.777706] env[63372]: DEBUG nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Refreshing aggregate associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, aggregates: None {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 691.799161] env[63372]: DEBUG nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Refreshing trait associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 691.988062] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Releasing lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.988062] env[63372]: DEBUG nova.compute.manager [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Instance network_info: |[{"id": "c8f78ebe-604d-4414-a16d-ae25256f2730", "address": "fa:16:3e:25:30:2b", "network": {"id": "73cbc616-354c-40de-971d-91fb5f0cbbac", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-817872371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2319173a3c445ca93ae2510141be32c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d", "external-id": "nsx-vlan-transportzone-591", "segmentation_id": 591, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8f78ebe-60", "ovs_interfaceid": "c8f78ebe-604d-4414-a16d-ae25256f2730", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 691.988062] env[63372]: DEBUG oslo_concurrency.lockutils [req-f2a26a71-5742-45cf-bf10-89d64496a94e req-25abf694-236b-4585-a023-59baeaedbaed service nova] Acquired lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.988062] env[63372]: DEBUG nova.network.neutron [req-f2a26a71-5742-45cf-bf10-89d64496a94e req-25abf694-236b-4585-a023-59baeaedbaed service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Refreshing network info cache for port c8f78ebe-604d-4414-a16d-ae25256f2730 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 691.989656] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:30:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c8f78ebe-604d-4414-a16d-ae25256f2730', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 691.999489] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Creating folder: Project (c2319173a3c445ca93ae2510141be32c). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 692.002895] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a84b7848-c05a-4945-b035-62fb969e402c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.020183] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Created folder: Project (c2319173a3c445ca93ae2510141be32c) in parent group-v227230. [ 692.020514] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Creating folder: Instances. Parent ref: group-v227280. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 692.020916] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7166b180-ef20-448f-8c24-e3bcd7b6b64d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.042019] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Created folder: Instances in parent group-v227280. [ 692.042019] env[63372]: DEBUG oslo.service.loopingcall [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 692.042019] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 692.042579] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-59eff706-a057-4a67-87ef-a40c3a2ab447 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.067790] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 692.067790] env[63372]: value = "task-1023800" [ 692.067790] env[63372]: _type = "Task" [ 692.067790] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.087582] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023800, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.135163] env[63372]: DEBUG nova.compute.manager [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 692.156629] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023796, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.256451] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d264dc01-3266-4c92-92f3-53f966fbbb7d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.265069] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c403e9-77cd-4291-8b37-5e0be879415f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.295880] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd0cbeb-14bb-49f2-a2dc-1ca7b1f0b0d3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.305068] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba6f49c-428a-48ac-b93f-c2292b83214f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.321831] env[63372]: DEBUG nova.compute.provider_tree [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 692.578424] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023800, 'name': CreateVM_Task, 'duration_secs': 0.354237} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.578609] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 692.579286] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.579734] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.579734] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 692.579983] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aee282d7-1f62-472a-ab45-0e0b8bd66ef2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.585840] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Waiting for the task: (returnval){ [ 692.585840] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5242f485-9289-abcf-37b3-19ef8776dfc4" [ 692.585840] env[63372]: _type = "Task" [ 692.585840] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.595691] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5242f485-9289-abcf-37b3-19ef8776dfc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.644691] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023796, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.537393} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.644691] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 276c3490-240a-499b-ba64-4b02510b169c/276c3490-240a-499b-ba64-4b02510b169c.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 692.644691] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 692.644691] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-683714d2-7091-4e8a-8920-04578546b2e0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.653814] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Waiting for the task: (returnval){ [ 692.653814] env[63372]: value = "task-1023801" [ 692.653814] env[63372]: _type = "Task" [ 692.653814] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.664982] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023801, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.668920] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.804204] env[63372]: DEBUG nova.network.neutron [req-f2a26a71-5742-45cf-bf10-89d64496a94e req-25abf694-236b-4585-a023-59baeaedbaed service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Updated VIF entry in instance network info cache for port c8f78ebe-604d-4414-a16d-ae25256f2730. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 692.804582] env[63372]: DEBUG nova.network.neutron [req-f2a26a71-5742-45cf-bf10-89d64496a94e req-25abf694-236b-4585-a023-59baeaedbaed service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Updating instance_info_cache with network_info: [{"id": "c8f78ebe-604d-4414-a16d-ae25256f2730", "address": "fa:16:3e:25:30:2b", "network": {"id": "73cbc616-354c-40de-971d-91fb5f0cbbac", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-817872371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2319173a3c445ca93ae2510141be32c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d", "external-id": "nsx-vlan-transportzone-591", "segmentation_id": 591, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8f78ebe-60", "ovs_interfaceid": "c8f78ebe-604d-4414-a16d-ae25256f2730", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.855095] env[63372]: DEBUG nova.scheduler.client.report [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 64 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 692.855375] env[63372]: DEBUG nova.compute.provider_tree [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 64 to 65 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 692.855555] env[63372]: DEBUG nova.compute.provider_tree [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 693.096339] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5242f485-9289-abcf-37b3-19ef8776dfc4, 'name': SearchDatastore_Task, 'duration_secs': 0.011614} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.096655] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.096883] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 693.097133] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.097279] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.097597] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 693.097713] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a4375dd-8630-45cd-8ebd-f8066ac8a433 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.106827] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 693.107033] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 693.107750] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10d76265-6ee5-4a38-8803-7fb0d031958a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.113515] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Waiting for the task: (returnval){ [ 693.113515] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]522b26b4-73e8-5915-a071-a4e5c99d6bbc" [ 693.113515] env[63372]: _type = "Task" [ 693.113515] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.121862] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522b26b4-73e8-5915-a071-a4e5c99d6bbc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.166510] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023801, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062412} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.166823] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 693.167576] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84cec12-4dfe-417d-81ae-1b0ce140004c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.190218] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] 276c3490-240a-499b-ba64-4b02510b169c/276c3490-240a-499b-ba64-4b02510b169c.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 693.190515] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-600c158c-0a5a-4144-97e5-371ed9440c1a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.212048] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Waiting for the task: (returnval){ [ 693.212048] env[63372]: value = "task-1023802" [ 693.212048] env[63372]: _type = "Task" [ 693.212048] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.220395] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023802, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.306888] env[63372]: DEBUG oslo_concurrency.lockutils [req-f2a26a71-5742-45cf-bf10-89d64496a94e req-25abf694-236b-4585-a023-59baeaedbaed service nova] Releasing lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.361112] env[63372]: DEBUG oslo_concurrency.lockutils [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 5.702s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.361736] env[63372]: DEBUG nova.compute.manager [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 693.364807] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.799s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.366246] env[63372]: INFO nova.compute.claims [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 693.624397] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522b26b4-73e8-5915-a071-a4e5c99d6bbc, 'name': SearchDatastore_Task, 'duration_secs': 0.009316} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.625180] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-944487e7-7e0b-4b00-8251-088244bb7db9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.630813] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Waiting for the task: (returnval){ [ 693.630813] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]526d8de2-db8c-4d99-c076-1c3562014fe8" [ 693.630813] env[63372]: _type = "Task" [ 693.630813] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.639099] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526d8de2-db8c-4d99-c076-1c3562014fe8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.722953] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023802, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.802055] env[63372]: DEBUG nova.compute.manager [req-612323f1-5e62-41ec-85f5-9d21ac716e09 req-e7092834-64b7-4db8-87e6-b40db52e9fd3 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Received event network-changed-e1486286-9147-4e5e-88c1-60893a3338dd {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 693.802277] env[63372]: DEBUG nova.compute.manager [req-612323f1-5e62-41ec-85f5-9d21ac716e09 req-e7092834-64b7-4db8-87e6-b40db52e9fd3 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Refreshing instance network info cache due to event network-changed-e1486286-9147-4e5e-88c1-60893a3338dd. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 693.802488] env[63372]: DEBUG oslo_concurrency.lockutils [req-612323f1-5e62-41ec-85f5-9d21ac716e09 req-e7092834-64b7-4db8-87e6-b40db52e9fd3 service nova] Acquiring lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.802627] env[63372]: DEBUG oslo_concurrency.lockutils [req-612323f1-5e62-41ec-85f5-9d21ac716e09 req-e7092834-64b7-4db8-87e6-b40db52e9fd3 service nova] Acquired lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.802880] env[63372]: DEBUG nova.network.neutron [req-612323f1-5e62-41ec-85f5-9d21ac716e09 req-e7092834-64b7-4db8-87e6-b40db52e9fd3 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Refreshing network info cache for port e1486286-9147-4e5e-88c1-60893a3338dd {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 693.871659] env[63372]: DEBUG nova.compute.utils [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.874930] env[63372]: DEBUG nova.compute.manager [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Not allocating networking since 'none' was specified. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 694.142766] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526d8de2-db8c-4d99-c076-1c3562014fe8, 'name': SearchDatastore_Task, 'duration_secs': 0.009739} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.143056] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.143340] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] d50919ac-3a0b-46ac-a837-ca3e6ad05173/d50919ac-3a0b-46ac-a837-ca3e6ad05173.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 694.143597] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d8893a3-0c6d-4a19-8ac9-260761fd9229 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.154011] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Waiting for the task: (returnval){ [ 694.154011] env[63372]: value = "task-1023804" [ 694.154011] env[63372]: _type = "Task" [ 694.154011] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.164440] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1023804, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.224594] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023802, 'name': ReconfigVM_Task, 'duration_secs': 0.829733} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.224981] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Reconfigured VM instance instance-00000025 to attach disk [datastore1] 276c3490-240a-499b-ba64-4b02510b169c/276c3490-240a-499b-ba64-4b02510b169c.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 694.225536] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-59959340-6cb1-44be-83f2-68ed9fd6110e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.233560] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Waiting for the task: (returnval){ [ 694.233560] env[63372]: value = "task-1023805" [ 694.233560] env[63372]: _type = "Task" [ 694.233560] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.375604] env[63372]: DEBUG nova.compute.manager [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 694.556293] env[63372]: DEBUG nova.network.neutron [req-612323f1-5e62-41ec-85f5-9d21ac716e09 req-e7092834-64b7-4db8-87e6-b40db52e9fd3 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updated VIF entry in instance network info cache for port e1486286-9147-4e5e-88c1-60893a3338dd. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 694.556759] env[63372]: DEBUG nova.network.neutron [req-612323f1-5e62-41ec-85f5-9d21ac716e09 req-e7092834-64b7-4db8-87e6-b40db52e9fd3 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance_info_cache with network_info: [{"id": "e1486286-9147-4e5e-88c1-60893a3338dd", "address": "fa:16:3e:6c:92:ac", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1486286-91", "ovs_interfaceid": "e1486286-9147-4e5e-88c1-60893a3338dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.669215] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1023804, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.749228] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023805, 'name': Rename_Task, 'duration_secs': 0.363317} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.752490] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 694.753086] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad1d1cea-eb3b-46d2-8d82-bd5f793ad2be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.763872] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Waiting for the task: (returnval){ [ 694.763872] env[63372]: value = "task-1023806" [ 694.763872] env[63372]: _type = "Task" [ 694.763872] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.776292] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023806, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.797572] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3afb99-b285-4972-83ba-8229eec67cb1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.809019] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee97d3c6-860d-4732-83ee-9c121879802c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.837499] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee4b621-bfe2-4303-8260-251289cfa171 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.846237] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383de04d-c4e6-4a23-a7d8-af73c3011d5e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.860027] env[63372]: DEBUG nova.compute.provider_tree [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 695.059374] env[63372]: DEBUG oslo_concurrency.lockutils [req-612323f1-5e62-41ec-85f5-9d21ac716e09 req-e7092834-64b7-4db8-87e6-b40db52e9fd3 service nova] Releasing lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.167960] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1023804, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.628391} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.168261] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] d50919ac-3a0b-46ac-a837-ca3e6ad05173/d50919ac-3a0b-46ac-a837-ca3e6ad05173.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 695.168500] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 695.168763] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-204bf8fa-de2f-402c-aecd-70c4daf1e7f7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.177102] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Waiting for the task: (returnval){ [ 695.177102] env[63372]: value = "task-1023807" [ 695.177102] env[63372]: _type = "Task" [ 695.177102] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.185010] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1023807, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.274820] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023806, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.386349] env[63372]: DEBUG nova.compute.manager [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 695.391165] env[63372]: DEBUG nova.scheduler.client.report [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 65 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 695.391557] env[63372]: DEBUG nova.compute.provider_tree [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 65 to 66 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 695.391819] env[63372]: DEBUG nova.compute.provider_tree [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 695.412490] env[63372]: DEBUG nova.virt.hardware [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 695.412731] env[63372]: DEBUG nova.virt.hardware [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 695.412887] env[63372]: DEBUG nova.virt.hardware [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 695.413702] env[63372]: DEBUG nova.virt.hardware [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 695.413702] env[63372]: DEBUG nova.virt.hardware [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 695.413702] env[63372]: DEBUG nova.virt.hardware [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 695.413702] env[63372]: DEBUG nova.virt.hardware [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 695.413929] env[63372]: DEBUG nova.virt.hardware [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 695.413929] env[63372]: DEBUG nova.virt.hardware [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 695.414164] env[63372]: DEBUG nova.virt.hardware [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 695.414392] env[63372]: DEBUG nova.virt.hardware [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 695.415733] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f33b2a-83a9-4f4e-93b8-87a403e84695 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.426537] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ba3338-4600-4cf7-80ed-a0b2c68c4116 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.443060] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Instance VIF info [] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 695.450045] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Creating folder: Project (a6d2092aabd74a8889b8afde9b613917). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 695.450045] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-10ee3429-b51d-4795-b3ae-be95b72a765a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.460715] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Created folder: Project (a6d2092aabd74a8889b8afde9b613917) in parent group-v227230. [ 695.460984] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Creating folder: Instances. Parent ref: group-v227283. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 695.461327] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c4ad0d1a-441e-4b80-b7b9-60bc9f63d645 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.471548] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Created folder: Instances in parent group-v227283. [ 695.471813] env[63372]: DEBUG oslo.service.loopingcall [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 695.471985] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 695.472208] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d06c0d21-6c45-4f7c-a2c6-5923e4642c00 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.490278] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 695.490278] env[63372]: value = "task-1023810" [ 695.490278] env[63372]: _type = "Task" [ 695.490278] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.498437] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023810, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.691024] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1023807, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102144} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.691024] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 695.691024] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-199b11e5-e1e5-4fba-a0cf-38d5b722259d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.719586] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] d50919ac-3a0b-46ac-a837-ca3e6ad05173/d50919ac-3a0b-46ac-a837-ca3e6ad05173.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 695.719960] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24d2f1c7-0871-4fee-884a-0a3ed769bee4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.741068] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Waiting for the task: (returnval){ [ 695.741068] env[63372]: value = "task-1023811" [ 695.741068] env[63372]: _type = "Task" [ 695.741068] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.750598] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1023811, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.775398] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023806, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.898831] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.899282] env[63372]: DEBUG nova.compute.manager [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 695.902063] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.828s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.902299] env[63372]: DEBUG nova.objects.instance [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lazy-loading 'resources' on Instance uuid 6d4502b7-b9df-4fef-8400-9ff07b820e2b {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 696.001335] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023810, 'name': CreateVM_Task, 'duration_secs': 0.489736} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.001483] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 696.001830] env[63372]: DEBUG oslo_concurrency.lockutils [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.001986] env[63372]: DEBUG oslo_concurrency.lockutils [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.002437] env[63372]: DEBUG oslo_concurrency.lockutils [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 696.002623] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ae20a2d-c217-47d6-8634-eb460ddb3622 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.007927] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Waiting for the task: (returnval){ [ 696.007927] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52ecaa4b-2aed-aede-3344-1cadb6a71c75" [ 696.007927] env[63372]: _type = "Task" [ 696.007927] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.017540] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ecaa4b-2aed-aede-3344-1cadb6a71c75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.253801] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1023811, 'name': ReconfigVM_Task, 'duration_secs': 0.346591} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.254162] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Reconfigured VM instance instance-00000026 to attach disk [datastore1] d50919ac-3a0b-46ac-a837-ca3e6ad05173/d50919ac-3a0b-46ac-a837-ca3e6ad05173.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 696.254774] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de6e2e2c-33f8-4ca8-b87f-996f77229fc8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.262100] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Waiting for the task: (returnval){ [ 696.262100] env[63372]: value = "task-1023812" [ 696.262100] env[63372]: _type = "Task" [ 696.262100] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.273805] env[63372]: DEBUG oslo_vmware.api [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023806, 'name': PowerOnVM_Task, 'duration_secs': 1.363231} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.277161] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 696.277449] env[63372]: INFO nova.compute.manager [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Took 9.14 seconds to spawn the instance on the hypervisor. [ 696.277559] env[63372]: DEBUG nova.compute.manager [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 696.277827] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1023812, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.278540] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa74c12-b3d9-4498-8f33-042aa9c4d30c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.405538] env[63372]: DEBUG nova.compute.utils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 696.411269] env[63372]: DEBUG nova.compute.manager [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 696.411269] env[63372]: DEBUG nova.network.neutron [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 696.467354] env[63372]: DEBUG nova.policy [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03ac5819dca2445cb89ae46fdf8979e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bd516c044b44ac3aabc56d7fd09f5ff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 696.519125] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ecaa4b-2aed-aede-3344-1cadb6a71c75, 'name': SearchDatastore_Task, 'duration_secs': 0.010824} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.519420] env[63372]: DEBUG oslo_concurrency.lockutils [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.519641] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 696.519884] env[63372]: DEBUG oslo_concurrency.lockutils [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.520118] env[63372]: DEBUG oslo_concurrency.lockutils [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.520377] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 696.520560] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-57b26e3c-0a16-4ad6-9782-8595f82fbb05 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.529614] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 696.529804] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 696.530535] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7c96f64-6982-4d67-b37c-da32d0445c2e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.536832] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Waiting for the task: (returnval){ [ 696.536832] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52a65b62-9077-1238-f5aa-41dccf71ec55" [ 696.536832] env[63372]: _type = "Task" [ 696.536832] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.544546] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a65b62-9077-1238-f5aa-41dccf71ec55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.745348] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5850206-b112-43c7-adc9-2690bd71ff4d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.748371] env[63372]: DEBUG nova.network.neutron [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Successfully created port: f47ddf2b-5ce1-46cc-aeac-18507b30f21e {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 696.756597] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2954c88-dc5a-4026-b853-2053e2f4d42b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.796007] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a298e14-8589-4d05-b5a3-16bfa1a8cf29 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.803919] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1023812, 'name': Rename_Task, 'duration_secs': 0.179747} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.804386] env[63372]: INFO nova.compute.manager [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Took 30.43 seconds to build instance. [ 696.805690] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 696.806113] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-324885d6-4dc9-4219-9326-d49cf4c748c8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.811457] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff2e37ce-d26d-4d46-90d4-fd13efb1d33d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.820248] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Waiting for the task: (returnval){ [ 696.820248] env[63372]: value = "task-1023813" [ 696.820248] env[63372]: _type = "Task" [ 696.820248] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.833376] env[63372]: DEBUG nova.compute.provider_tree [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.837821] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1023813, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.911420] env[63372]: DEBUG nova.compute.manager [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 697.049327] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a65b62-9077-1238-f5aa-41dccf71ec55, 'name': SearchDatastore_Task, 'duration_secs': 0.008983} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.050439] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c3fa007-42dd-42c8-9258-29a269ca88b9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.056883] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Waiting for the task: (returnval){ [ 697.056883] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f887f9-8c45-a923-c0dc-c26bec7aa295" [ 697.056883] env[63372]: _type = "Task" [ 697.056883] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.065753] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f887f9-8c45-a923-c0dc-c26bec7aa295, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.307027] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a5c31a6-9d78-48f8-be02-ceb0f1abff5b tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Lock "276c3490-240a-499b-ba64-4b02510b169c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.683s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.332842] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1023813, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.339956] env[63372]: DEBUG nova.scheduler.client.report [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.568587] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f887f9-8c45-a923-c0dc-c26bec7aa295, 'name': SearchDatastore_Task, 'duration_secs': 0.019895} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.568800] env[63372]: DEBUG oslo_concurrency.lockutils [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.569058] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 44fba254-7dec-4458-9ae3-fdbbe4895de5/44fba254-7dec-4458-9ae3-fdbbe4895de5.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 697.569348] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79e297d4-3179-46b6-b67a-e8ba44441ef9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.577574] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Waiting for the task: (returnval){ [ 697.577574] env[63372]: value = "task-1023814" [ 697.577574] env[63372]: _type = "Task" [ 697.577574] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.587387] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': task-1023814, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.809577] env[63372]: DEBUG nova.compute.manager [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 697.835372] env[63372]: DEBUG oslo_vmware.api [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1023813, 'name': PowerOnVM_Task, 'duration_secs': 1.000912} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.835788] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 697.836153] env[63372]: INFO nova.compute.manager [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Took 8.15 seconds to spawn the instance on the hypervisor. [ 697.836398] env[63372]: DEBUG nova.compute.manager [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 697.837560] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52dfb224-49b0-4489-8e4c-27c0edb3bdb4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.847429] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.945s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.853143] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.728s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.854926] env[63372]: INFO nova.compute.claims [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.870791] env[63372]: INFO nova.scheduler.client.report [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Deleted allocations for instance 6d4502b7-b9df-4fef-8400-9ff07b820e2b [ 697.921282] env[63372]: DEBUG nova.compute.manager [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 697.955573] env[63372]: DEBUG nova.virt.hardware [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 697.955882] env[63372]: DEBUG nova.virt.hardware [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 697.956115] env[63372]: DEBUG nova.virt.hardware [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 697.956328] env[63372]: DEBUG nova.virt.hardware [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 697.956476] env[63372]: DEBUG nova.virt.hardware [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 697.956623] env[63372]: DEBUG nova.virt.hardware [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 697.956832] env[63372]: DEBUG nova.virt.hardware [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 697.957029] env[63372]: DEBUG nova.virt.hardware [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 697.957262] env[63372]: DEBUG nova.virt.hardware [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 697.957480] env[63372]: DEBUG nova.virt.hardware [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 697.957720] env[63372]: DEBUG nova.virt.hardware [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 697.958719] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dabeabc8-ee0d-4f7a-a30d-c3c5b362e3bd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.968576] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb827b59-cb4e-4fa8-8297-f07b5f91357e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.088786] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': task-1023814, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508785} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.089072] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 44fba254-7dec-4458-9ae3-fdbbe4895de5/44fba254-7dec-4458-9ae3-fdbbe4895de5.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 698.089288] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 698.089604] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f978002b-04ce-4140-8680-bc72bbb4947e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.093437] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Acquiring lock "276c3490-240a-499b-ba64-4b02510b169c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.093720] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Lock "276c3490-240a-499b-ba64-4b02510b169c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.093962] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Acquiring lock "276c3490-240a-499b-ba64-4b02510b169c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.094180] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Lock "276c3490-240a-499b-ba64-4b02510b169c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.094498] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Lock "276c3490-240a-499b-ba64-4b02510b169c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.097356] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Waiting for the task: (returnval){ [ 698.097356] env[63372]: value = "task-1023815" [ 698.097356] env[63372]: _type = "Task" [ 698.097356] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.097731] env[63372]: INFO nova.compute.manager [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Terminating instance [ 698.103204] env[63372]: DEBUG nova.compute.manager [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 698.103204] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 698.103634] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cec5ee9-d291-492f-820a-9579a63547c7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.111993] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': task-1023815, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.114947] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 698.115247] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6873b6fc-0695-444b-82bd-e6f7e7db2ec5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.124087] env[63372]: DEBUG oslo_vmware.api [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Waiting for the task: (returnval){ [ 698.124087] env[63372]: value = "task-1023816" [ 698.124087] env[63372]: _type = "Task" [ 698.124087] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.134463] env[63372]: DEBUG oslo_vmware.api [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023816, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.241759] env[63372]: DEBUG nova.compute.manager [req-613d4071-c1b4-4b1e-bcc8-1caabe452400 req-3f53603d-6253-415d-af9d-d38e3e848dea service nova] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Received event network-vif-plugged-f47ddf2b-5ce1-46cc-aeac-18507b30f21e {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 698.241985] env[63372]: DEBUG oslo_concurrency.lockutils [req-613d4071-c1b4-4b1e-bcc8-1caabe452400 req-3f53603d-6253-415d-af9d-d38e3e848dea service nova] Acquiring lock "85883c62-6d5e-4fa6-b4f0-afac984dc6bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.242203] env[63372]: DEBUG oslo_concurrency.lockutils [req-613d4071-c1b4-4b1e-bcc8-1caabe452400 req-3f53603d-6253-415d-af9d-d38e3e848dea service nova] Lock "85883c62-6d5e-4fa6-b4f0-afac984dc6bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.242369] env[63372]: DEBUG oslo_concurrency.lockutils [req-613d4071-c1b4-4b1e-bcc8-1caabe452400 req-3f53603d-6253-415d-af9d-d38e3e848dea service nova] Lock "85883c62-6d5e-4fa6-b4f0-afac984dc6bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.242531] env[63372]: DEBUG nova.compute.manager [req-613d4071-c1b4-4b1e-bcc8-1caabe452400 req-3f53603d-6253-415d-af9d-d38e3e848dea service nova] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] No waiting events found dispatching network-vif-plugged-f47ddf2b-5ce1-46cc-aeac-18507b30f21e {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 698.242689] env[63372]: WARNING nova.compute.manager [req-613d4071-c1b4-4b1e-bcc8-1caabe452400 req-3f53603d-6253-415d-af9d-d38e3e848dea service nova] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Received unexpected event network-vif-plugged-f47ddf2b-5ce1-46cc-aeac-18507b30f21e for instance with vm_state building and task_state spawning. [ 698.333568] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.352136] env[63372]: DEBUG nova.network.neutron [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Successfully updated port: f47ddf2b-5ce1-46cc-aeac-18507b30f21e {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 698.374192] env[63372]: INFO nova.compute.manager [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Took 30.47 seconds to build instance. [ 698.375375] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Acquiring lock "69a107d0-80c7-42e5-b514-b4273e1a3359" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.375591] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Lock "69a107d0-80c7-42e5-b514-b4273e1a3359" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.382685] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbdd6a0e-3865-47be-8bb6-d4aa665a96f4 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "6d4502b7-b9df-4fef-8400-9ff07b820e2b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.258s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.615040] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': task-1023815, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064733} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.615040] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 698.615040] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48637cfb-1b20-41f9-b503-36839b211a99 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.635720] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] 44fba254-7dec-4458-9ae3-fdbbe4895de5/44fba254-7dec-4458-9ae3-fdbbe4895de5.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 698.639320] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6dbc5888-9b2f-45bc-9cc8-e1dadd257a94 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.660045] env[63372]: DEBUG oslo_vmware.api [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023816, 'name': PowerOffVM_Task, 'duration_secs': 0.297047} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.661567] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 698.661892] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 698.662302] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Waiting for the task: (returnval){ [ 698.662302] env[63372]: value = "task-1023817" [ 698.662302] env[63372]: _type = "Task" [ 698.662302] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.662689] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-350c41d6-8c59-4a62-bde8-830a6200fb66 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.672873] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': task-1023817, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.756067] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 698.756067] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 698.756067] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Deleting the datastore file [datastore1] 276c3490-240a-499b-ba64-4b02510b169c {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 698.756067] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4fcaf346-d741-4cd7-9735-097615923fe5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.768162] env[63372]: DEBUG oslo_vmware.api [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Waiting for the task: (returnval){ [ 698.768162] env[63372]: value = "task-1023819" [ 698.768162] env[63372]: _type = "Task" [ 698.768162] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.778181] env[63372]: DEBUG oslo_vmware.api [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023819, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.856292] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Acquiring lock "refresh_cache-85883c62-6d5e-4fa6-b4f0-afac984dc6bb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.856754] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Acquired lock "refresh_cache-85883c62-6d5e-4fa6-b4f0-afac984dc6bb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.856974] env[63372]: DEBUG nova.network.neutron [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 698.879406] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3205410a-fd27-4889-a24a-67feb7b0def1 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Lock "d50919ac-3a0b-46ac-a837-ca3e6ad05173" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.067s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.180319] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': task-1023817, 'name': ReconfigVM_Task, 'duration_secs': 0.428897} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.184245] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Reconfigured VM instance instance-00000027 to attach disk [datastore2] 44fba254-7dec-4458-9ae3-fdbbe4895de5/44fba254-7dec-4458-9ae3-fdbbe4895de5.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 699.184303] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-829deadc-509d-4e97-bd19-b881adc091a6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.194259] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Waiting for the task: (returnval){ [ 699.194259] env[63372]: value = "task-1023820" [ 699.194259] env[63372]: _type = "Task" [ 699.194259] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.208170] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': task-1023820, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.278909] env[63372]: DEBUG oslo_vmware.api [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Task: {'id': task-1023819, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15225} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.278909] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 699.279064] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 699.279167] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 699.279334] env[63372]: INFO nova.compute.manager [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Took 1.18 seconds to destroy the instance on the hypervisor. [ 699.279570] env[63372]: DEBUG oslo.service.loopingcall [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 699.283083] env[63372]: DEBUG nova.compute.manager [-] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 699.283083] env[63372]: DEBUG nova.network.neutron [-] [instance: 276c3490-240a-499b-ba64-4b02510b169c] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 699.287545] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b25ab04d-9a87-4384-b1c6-5a666e8c0028 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.296275] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf41545-89e4-4fa7-8d51-04e43374074f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.335361] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ecdd53-5935-4cbe-85a3-bcf53fefb841 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.345792] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5691d54-5189-4ef9-9c52-097490263ba5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.363587] env[63372]: DEBUG nova.compute.provider_tree [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 699.382038] env[63372]: DEBUG nova.compute.manager [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 699.431969] env[63372]: DEBUG nova.network.neutron [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.645549] env[63372]: DEBUG nova.network.neutron [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Updating instance_info_cache with network_info: [{"id": "f47ddf2b-5ce1-46cc-aeac-18507b30f21e", "address": "fa:16:3e:5f:b8:bf", "network": {"id": "3c2e219f-e82f-47f6-8038-9148d249660c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1634022873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bd516c044b44ac3aabc56d7fd09f5ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf47ddf2b-5c", "ovs_interfaceid": "f47ddf2b-5ce1-46cc-aeac-18507b30f21e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.708269] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': task-1023820, 'name': Rename_Task, 'duration_secs': 0.151783} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.708508] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 699.709015] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4c192d4c-6b96-4cbc-b2fb-8efe0d0ab35c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.717444] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Waiting for the task: (returnval){ [ 699.717444] env[63372]: value = "task-1023821" [ 699.717444] env[63372]: _type = "Task" [ 699.717444] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.725478] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': task-1023821, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.887770] env[63372]: ERROR nova.scheduler.client.report [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [req-8a47cd78-7f42-44aa-99ae-47c78d216e09] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8a47cd78-7f42-44aa-99ae-47c78d216e09"}]} [ 699.907754] env[63372]: DEBUG nova.scheduler.client.report [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Refreshing inventories for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 699.918012] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.920616] env[63372]: DEBUG nova.scheduler.client.report [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Updating ProviderTree inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 699.920811] env[63372]: DEBUG nova.compute.provider_tree [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 699.931167] env[63372]: DEBUG nova.scheduler.client.report [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Refreshing aggregate associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, aggregates: None {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 699.950435] env[63372]: DEBUG nova.scheduler.client.report [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Refreshing trait associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 700.149673] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Releasing lock "refresh_cache-85883c62-6d5e-4fa6-b4f0-afac984dc6bb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.150028] env[63372]: DEBUG nova.compute.manager [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Instance network_info: |[{"id": "f47ddf2b-5ce1-46cc-aeac-18507b30f21e", "address": "fa:16:3e:5f:b8:bf", "network": {"id": "3c2e219f-e82f-47f6-8038-9148d249660c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1634022873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bd516c044b44ac3aabc56d7fd09f5ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf47ddf2b-5c", "ovs_interfaceid": "f47ddf2b-5ce1-46cc-aeac-18507b30f21e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 700.150473] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5f:b8:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca50cd14-9e1f-4d74-a066-e5a45ba0ce22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f47ddf2b-5ce1-46cc-aeac-18507b30f21e', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 700.159306] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Creating folder: Project (7bd516c044b44ac3aabc56d7fd09f5ff). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 700.162336] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ae04f518-9fd8-4b2c-b557-b3bea0cabad4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.176207] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Created folder: Project (7bd516c044b44ac3aabc56d7fd09f5ff) in parent group-v227230. [ 700.176452] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Creating folder: Instances. Parent ref: group-v227286. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 700.176712] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-acc9aff1-b126-4633-8710-d9da3f702048 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.191578] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Created folder: Instances in parent group-v227286. [ 700.191841] env[63372]: DEBUG oslo.service.loopingcall [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.192061] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 700.192248] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48bae2a9-ba9c-45ae-bbd4-f9c306a8c97f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.209170] env[63372]: DEBUG nova.network.neutron [-] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.216854] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 700.216854] env[63372]: value = "task-1023824" [ 700.216854] env[63372]: _type = "Task" [ 700.216854] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.230645] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023824, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.239021] env[63372]: DEBUG oslo_vmware.api [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': task-1023821, 'name': PowerOnVM_Task, 'duration_secs': 0.485211} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.239021] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 700.239021] env[63372]: INFO nova.compute.manager [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Took 4.85 seconds to spawn the instance on the hypervisor. [ 700.239021] env[63372]: DEBUG nova.compute.manager [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 700.239021] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f015c8-a47b-4da9-a761-b93cb19f5a81 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.287840] env[63372]: DEBUG nova.compute.manager [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Received event network-changed-f47ddf2b-5ce1-46cc-aeac-18507b30f21e {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 700.288106] env[63372]: DEBUG nova.compute.manager [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Refreshing instance network info cache due to event network-changed-f47ddf2b-5ce1-46cc-aeac-18507b30f21e. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 700.288346] env[63372]: DEBUG oslo_concurrency.lockutils [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] Acquiring lock "refresh_cache-85883c62-6d5e-4fa6-b4f0-afac984dc6bb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.288564] env[63372]: DEBUG oslo_concurrency.lockutils [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] Acquired lock "refresh_cache-85883c62-6d5e-4fa6-b4f0-afac984dc6bb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.288761] env[63372]: DEBUG nova.network.neutron [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Refreshing network info cache for port f47ddf2b-5ce1-46cc-aeac-18507b30f21e {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 700.399598] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8894d6e0-e43e-4cb1-8fb2-7cbe021c2b20 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.408767] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908d95b5-faa0-4034-b0b9-295cc8270512 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.440965] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2294dc96-2b5c-4ab5-ab5a-d72299c55751 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.450224] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64490464-116f-417a-9ac1-ad861f1c78ad {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.466238] env[63372]: DEBUG nova.compute.provider_tree [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 700.712224] env[63372]: INFO nova.compute.manager [-] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Took 1.43 seconds to deallocate network for instance. [ 700.727811] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023824, 'name': CreateVM_Task, 'duration_secs': 0.349242} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.727978] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 700.728690] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.728854] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.729186] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 700.729430] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41eebd5c-abda-4440-9cac-e06507f1d6b6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.735203] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Waiting for the task: (returnval){ [ 700.735203] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52aedf89-20a0-c6e2-f780-d3ed6522ccb2" [ 700.735203] env[63372]: _type = "Task" [ 700.735203] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.753376] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52aedf89-20a0-c6e2-f780-d3ed6522ccb2, 'name': SearchDatastore_Task, 'duration_secs': 0.010498} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.759731] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.760419] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 700.760419] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.760419] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.760664] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 700.762824] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c5fccccd-e941-4457-bdbd-137aedd87fc7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.764887] env[63372]: INFO nova.compute.manager [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Took 32.74 seconds to build instance. [ 700.775178] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 700.775178] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 700.776215] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-807e94fe-174c-4d3a-9e11-5bc153ee45a9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.784051] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Waiting for the task: (returnval){ [ 700.784051] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5297815a-b8b9-a1df-4246-3317571c65a6" [ 700.784051] env[63372]: _type = "Task" [ 700.784051] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.798174] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5297815a-b8b9-a1df-4246-3317571c65a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.005560] env[63372]: DEBUG nova.scheduler.client.report [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 67 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 701.005887] env[63372]: DEBUG nova.compute.provider_tree [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 67 to 68 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 701.006012] env[63372]: DEBUG nova.compute.provider_tree [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 701.022341] env[63372]: DEBUG nova.network.neutron [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Updated VIF entry in instance network info cache for port f47ddf2b-5ce1-46cc-aeac-18507b30f21e. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 701.022733] env[63372]: DEBUG nova.network.neutron [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Updating instance_info_cache with network_info: [{"id": "f47ddf2b-5ce1-46cc-aeac-18507b30f21e", "address": "fa:16:3e:5f:b8:bf", "network": {"id": "3c2e219f-e82f-47f6-8038-9148d249660c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1634022873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bd516c044b44ac3aabc56d7fd09f5ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf47ddf2b-5c", "ovs_interfaceid": "f47ddf2b-5ce1-46cc-aeac-18507b30f21e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.218826] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.267283] env[63372]: DEBUG oslo_concurrency.lockutils [None req-467894e0-59b7-4dba-b14d-68416733c4a1 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Lock "44fba254-7dec-4458-9ae3-fdbbe4895de5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.870s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.295495] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5297815a-b8b9-a1df-4246-3317571c65a6, 'name': SearchDatastore_Task, 'duration_secs': 0.010063} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.296562] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ff33b45-3fae-4009-9371-795b4d78307f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.302934] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Waiting for the task: (returnval){ [ 701.302934] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fb0913-93c7-b28b-097d-9add2c445a74" [ 701.302934] env[63372]: _type = "Task" [ 701.302934] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.311643] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fb0913-93c7-b28b-097d-9add2c445a74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.514140] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.659s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.514140] env[63372]: DEBUG nova.compute.manager [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 701.516868] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.069s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.518648] env[63372]: INFO nova.compute.claims [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.524829] env[63372]: DEBUG oslo_concurrency.lockutils [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] Releasing lock "refresh_cache-85883c62-6d5e-4fa6-b4f0-afac984dc6bb" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.525093] env[63372]: DEBUG nova.compute.manager [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Received event network-changed-c8f78ebe-604d-4414-a16d-ae25256f2730 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.525236] env[63372]: DEBUG nova.compute.manager [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Refreshing instance network info cache due to event network-changed-c8f78ebe-604d-4414-a16d-ae25256f2730. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 701.525468] env[63372]: DEBUG oslo_concurrency.lockutils [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] Acquiring lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.525614] env[63372]: DEBUG oslo_concurrency.lockutils [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] Acquired lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.525773] env[63372]: DEBUG nova.network.neutron [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Refreshing network info cache for port c8f78ebe-604d-4414-a16d-ae25256f2730 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 701.649331] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "febeca1c-3370-4c74-8bd9-efad313df1e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.649621] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "febeca1c-3370-4c74-8bd9-efad313df1e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.770044] env[63372]: DEBUG nova.compute.manager [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 701.816036] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fb0913-93c7-b28b-097d-9add2c445a74, 'name': SearchDatastore_Task, 'duration_secs': 0.02899} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.816372] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.816629] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 85883c62-6d5e-4fa6-b4f0-afac984dc6bb/85883c62-6d5e-4fa6-b4f0-afac984dc6bb.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 701.816921] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a6b60fd-a5d0-44b2-ba74-73f2433a96a5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.823686] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Waiting for the task: (returnval){ [ 701.823686] env[63372]: value = "task-1023825" [ 701.823686] env[63372]: _type = "Task" [ 701.823686] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.832042] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023825, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.024033] env[63372]: DEBUG nova.compute.utils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 702.028066] env[63372]: DEBUG nova.compute.manager [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 702.028066] env[63372]: DEBUG nova.network.neutron [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 702.117020] env[63372]: DEBUG nova.policy [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '855ce9d887cd463cbf1a0f74e1197808', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b68c21179fb24008baf1da0d0119fb29', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 702.133449] env[63372]: DEBUG nova.compute.manager [None req-b0a8f31c-3782-45b6-b839-9289e5f16bd9 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 702.134440] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0972903-b31a-49ee-afdb-83de37d9f460 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.232273] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Acquiring lock "44fba254-7dec-4458-9ae3-fdbbe4895de5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.232829] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Lock "44fba254-7dec-4458-9ae3-fdbbe4895de5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.233105] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Acquiring lock "44fba254-7dec-4458-9ae3-fdbbe4895de5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.233354] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Lock "44fba254-7dec-4458-9ae3-fdbbe4895de5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.233722] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Lock "44fba254-7dec-4458-9ae3-fdbbe4895de5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.236698] env[63372]: INFO nova.compute.manager [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Terminating instance [ 702.241950] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Acquiring lock "refresh_cache-44fba254-7dec-4458-9ae3-fdbbe4895de5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.242131] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Acquired lock "refresh_cache-44fba254-7dec-4458-9ae3-fdbbe4895de5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.242311] env[63372]: DEBUG nova.network.neutron [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 702.290907] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 702.291226] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 702.291294] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Starting heal instance info cache {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 702.291377] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Rebuilding the list of instances to heal {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 702.299730] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.334553] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023825, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464109} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.334818] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 85883c62-6d5e-4fa6-b4f0-afac984dc6bb/85883c62-6d5e-4fa6-b4f0-afac984dc6bb.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 702.335042] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 702.335293] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f431d390-71d4-47be-9d16-a8c910dfd646 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.343527] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Waiting for the task: (returnval){ [ 702.343527] env[63372]: value = "task-1023826" [ 702.343527] env[63372]: _type = "Task" [ 702.343527] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.352997] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023826, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.529452] env[63372]: DEBUG nova.compute.manager [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 702.601642] env[63372]: DEBUG nova.network.neutron [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Updated VIF entry in instance network info cache for port c8f78ebe-604d-4414-a16d-ae25256f2730. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 702.605019] env[63372]: DEBUG nova.network.neutron [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Updating instance_info_cache with network_info: [{"id": "c8f78ebe-604d-4414-a16d-ae25256f2730", "address": "fa:16:3e:25:30:2b", "network": {"id": "73cbc616-354c-40de-971d-91fb5f0cbbac", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-817872371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2319173a3c445ca93ae2510141be32c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d", "external-id": "nsx-vlan-transportzone-591", "segmentation_id": 591, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8f78ebe-60", "ovs_interfaceid": "c8f78ebe-604d-4414-a16d-ae25256f2730", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.649189] env[63372]: INFO nova.compute.manager [None req-b0a8f31c-3782-45b6-b839-9289e5f16bd9 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] instance snapshotting [ 702.649873] env[63372]: DEBUG nova.objects.instance [None req-b0a8f31c-3782-45b6-b839-9289e5f16bd9 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Lazy-loading 'flavor' on Instance uuid 44fba254-7dec-4458-9ae3-fdbbe4895de5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 702.681884] env[63372]: DEBUG nova.network.neutron [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Successfully created port: 7688e644-e47b-4992-b037-343a71e18d2f {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.767405] env[63372]: DEBUG nova.network.neutron [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.802128] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Skipping network cache update for instance because it is being deleted. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 702.802128] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 702.802128] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 702.802732] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 702.843059] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "refresh_cache-cf673ac1-2c7d-468b-83ec-c723d5182457" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.843200] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquired lock "refresh_cache-cf673ac1-2c7d-468b-83ec-c723d5182457" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.843329] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Forcefully refreshing network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 702.843482] env[63372]: DEBUG nova.objects.instance [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lazy-loading 'info_cache' on Instance uuid cf673ac1-2c7d-468b-83ec-c723d5182457 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 702.853130] env[63372]: DEBUG nova.network.neutron [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.859762] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023826, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081634} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.860372] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 702.866506] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c993edb-5660-45e3-8ed8-40a53bc88496 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.893362] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 85883c62-6d5e-4fa6-b4f0-afac984dc6bb/85883c62-6d5e-4fa6-b4f0-afac984dc6bb.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 702.898348] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6964d1c4-aab2-402b-af09-9b159aa85de2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.922369] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Waiting for the task: (returnval){ [ 702.922369] env[63372]: value = "task-1023827" [ 702.922369] env[63372]: _type = "Task" [ 702.922369] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.933600] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023827, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.026048] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb5ffe2-f3bb-41b8-973b-c2f271d199fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.035854] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9986bd44-0ad2-4e85-959c-0a829d384a17 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.070808] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01921704-c12e-4d8b-9168-f2e38328ebc8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.078944] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a212c2c-b132-415e-acec-701d2a0211c3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.093482] env[63372]: DEBUG nova.compute.provider_tree [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.104760] env[63372]: DEBUG oslo_concurrency.lockutils [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] Releasing lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.105426] env[63372]: DEBUG nova.compute.manager [req-42931257-85ba-4f38-99e8-c618b59a4e4b req-1d57698f-0bc6-48fb-9486-0d605e78ee12 service nova] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Received event network-vif-deleted-7cb73607-f39d-49bc-ac64-cec7fd400af3 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 703.160691] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741b6937-01c8-4e6d-929e-24fceee760fd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.182026] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cc7f39e-6af7-443d-a58e-aee6330119c0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.361605] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Releasing lock "refresh_cache-44fba254-7dec-4458-9ae3-fdbbe4895de5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.361834] env[63372]: DEBUG nova.compute.manager [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 703.361867] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 703.362935] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb121bf-b7ae-4cb2-8dbb-673ff54cb014 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.373335] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 703.373780] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-68cdc9ad-7882-4b34-ae21-533e1ed64b37 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.382492] env[63372]: DEBUG oslo_vmware.api [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Waiting for the task: (returnval){ [ 703.382492] env[63372]: value = "task-1023828" [ 703.382492] env[63372]: _type = "Task" [ 703.382492] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.402025] env[63372]: DEBUG oslo_vmware.api [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': task-1023828, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.433144] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023827, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.544132] env[63372]: DEBUG nova.compute.manager [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 703.574320] env[63372]: DEBUG nova.virt.hardware [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 703.574919] env[63372]: DEBUG nova.virt.hardware [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 703.575341] env[63372]: DEBUG nova.virt.hardware [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.575614] env[63372]: DEBUG nova.virt.hardware [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 703.575781] env[63372]: DEBUG nova.virt.hardware [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.575935] env[63372]: DEBUG nova.virt.hardware [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 703.576285] env[63372]: DEBUG nova.virt.hardware [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 703.576409] env[63372]: DEBUG nova.virt.hardware [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 703.576582] env[63372]: DEBUG nova.virt.hardware [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 703.576742] env[63372]: DEBUG nova.virt.hardware [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 703.576910] env[63372]: DEBUG nova.virt.hardware [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 703.577819] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d7c1a2-131e-4f60-917d-8b04e0ca4176 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.587695] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c93b41-541b-4d05-9923-fb27f3b18fcb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.598682] env[63372]: DEBUG nova.scheduler.client.report [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.695714] env[63372]: DEBUG nova.compute.manager [None req-b0a8f31c-3782-45b6-b839-9289e5f16bd9 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Instance disappeared during snapshot {{(pid=63372) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 703.823533] env[63372]: DEBUG nova.compute.manager [None req-b0a8f31c-3782-45b6-b839-9289e5f16bd9 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Found 0 images (rotation: 2) {{(pid=63372) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 703.893979] env[63372]: DEBUG oslo_vmware.api [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': task-1023828, 'name': PowerOffVM_Task, 'duration_secs': 0.155079} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.894296] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 703.894459] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 703.894693] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-db3e9cdd-046e-4bb1-a9a0-3707a1055835 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.924885] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 703.925186] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 703.925402] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Deleting the datastore file [datastore2] 44fba254-7dec-4458-9ae3-fdbbe4895de5 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 703.929703] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bdeb0b1f-8d42-4ce9-842e-b31604aac865 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.936344] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023827, 'name': ReconfigVM_Task, 'duration_secs': 0.53781} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.937688] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 85883c62-6d5e-4fa6-b4f0-afac984dc6bb/85883c62-6d5e-4fa6-b4f0-afac984dc6bb.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 703.938448] env[63372]: DEBUG oslo_vmware.api [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Waiting for the task: (returnval){ [ 703.938448] env[63372]: value = "task-1023830" [ 703.938448] env[63372]: _type = "Task" [ 703.938448] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.938640] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-116c5966-a29c-411a-89dc-83b436940b6b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.948831] env[63372]: DEBUG oslo_vmware.api [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': task-1023830, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.950256] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Waiting for the task: (returnval){ [ 703.950256] env[63372]: value = "task-1023831" [ 703.950256] env[63372]: _type = "Task" [ 703.950256] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.959251] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023831, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.111999] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.595s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.112886] env[63372]: DEBUG nova.compute.manager [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 704.115385] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.618s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.115625] env[63372]: DEBUG nova.objects.instance [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Lazy-loading 'resources' on Instance uuid 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 704.170355] env[63372]: DEBUG nova.compute.manager [req-d61e42b8-c36c-468c-b174-35eb2f1a1ecb req-5504f985-2c80-48e5-ab7b-2dc6ada6cd7f service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Received event network-vif-plugged-7688e644-e47b-4992-b037-343a71e18d2f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.170355] env[63372]: DEBUG oslo_concurrency.lockutils [req-d61e42b8-c36c-468c-b174-35eb2f1a1ecb req-5504f985-2c80-48e5-ab7b-2dc6ada6cd7f service nova] Acquiring lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.170644] env[63372]: DEBUG oslo_concurrency.lockutils [req-d61e42b8-c36c-468c-b174-35eb2f1a1ecb req-5504f985-2c80-48e5-ab7b-2dc6ada6cd7f service nova] Lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.170715] env[63372]: DEBUG oslo_concurrency.lockutils [req-d61e42b8-c36c-468c-b174-35eb2f1a1ecb req-5504f985-2c80-48e5-ab7b-2dc6ada6cd7f service nova] Lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.170866] env[63372]: DEBUG nova.compute.manager [req-d61e42b8-c36c-468c-b174-35eb2f1a1ecb req-5504f985-2c80-48e5-ab7b-2dc6ada6cd7f service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] No waiting events found dispatching network-vif-plugged-7688e644-e47b-4992-b037-343a71e18d2f {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 704.171039] env[63372]: WARNING nova.compute.manager [req-d61e42b8-c36c-468c-b174-35eb2f1a1ecb req-5504f985-2c80-48e5-ab7b-2dc6ada6cd7f service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Received unexpected event network-vif-plugged-7688e644-e47b-4992-b037-343a71e18d2f for instance with vm_state building and task_state spawning. [ 704.268966] env[63372]: DEBUG nova.network.neutron [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Successfully updated port: 7688e644-e47b-4992-b037-343a71e18d2f {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 704.450994] env[63372]: DEBUG oslo_vmware.api [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Task: {'id': task-1023830, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134566} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.451447] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 704.451640] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 704.451864] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 704.452064] env[63372]: INFO nova.compute.manager [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Took 1.09 seconds to destroy the instance on the hypervisor. [ 704.452291] env[63372]: DEBUG oslo.service.loopingcall [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 704.455460] env[63372]: DEBUG nova.compute.manager [-] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.455573] env[63372]: DEBUG nova.network.neutron [-] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 704.462789] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023831, 'name': Rename_Task, 'duration_secs': 0.172157} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.463079] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 704.463326] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ae629a7b-a629-4c74-be44-4900c02d6696 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.471423] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Waiting for the task: (returnval){ [ 704.471423] env[63372]: value = "task-1023832" [ 704.471423] env[63372]: _type = "Task" [ 704.471423] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.479274] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023832, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.481618] env[63372]: DEBUG nova.network.neutron [-] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.590730] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Updating instance_info_cache with network_info: [{"id": "e56add8e-4342-4ab4-b2f2-c03f257f3d3a", "address": "fa:16:3e:11:78:8c", "network": {"id": "79050166-5dc7-4ef3-9819-ac534e041985", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-2103555585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06cecd1557de407697ad403a85138481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24727047-6358-4015-86c1-394ab07fb88f", "external-id": "nsx-vlan-transportzone-476", "segmentation_id": 476, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape56add8e-43", "ovs_interfaceid": "e56add8e-4342-4ab4-b2f2-c03f257f3d3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.618934] env[63372]: DEBUG nova.compute.utils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.625135] env[63372]: DEBUG nova.compute.manager [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.625135] env[63372]: DEBUG nova.network.neutron [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 704.667696] env[63372]: DEBUG nova.policy [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '335b72fcebbf4d62a43ac256b57664e7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b971b7d7a62d4902a930b10d96cfb51e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 704.771597] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.771748] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquired lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.771896] env[63372]: DEBUG nova.network.neutron [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 704.923645] env[63372]: DEBUG nova.network.neutron [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Successfully created port: 620d14ec-96a7-4912-bb91-44d2c290d431 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 704.980114] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c2df76-54eb-4b5e-9264-62e46e180434 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.984695] env[63372]: DEBUG nova.network.neutron [-] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.986052] env[63372]: DEBUG oslo_vmware.api [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023832, 'name': PowerOnVM_Task, 'duration_secs': 0.424542} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.990053] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 704.990053] env[63372]: INFO nova.compute.manager [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Took 7.07 seconds to spawn the instance on the hypervisor. [ 704.990053] env[63372]: DEBUG nova.compute.manager [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 704.990053] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0e4a15-b9df-4381-b151-1521dc9493c9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.994718] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7baba572-0a09-4c02-8dbe-cb293779cfe1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.032352] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728e871c-25cf-4bf7-9aae-2cb12af763e6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.042238] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4459b59-f329-486e-a79c-ae842ea866b8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.057687] env[63372]: DEBUG nova.compute.provider_tree [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.095289] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Releasing lock "refresh_cache-cf673ac1-2c7d-468b-83ec-c723d5182457" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.095503] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Updated the network info_cache for instance {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 705.095701] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 705.095858] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 705.096011] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 705.096250] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 705.096422] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 705.096570] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 705.096695] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63372) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 705.096836] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager.update_available_resource {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 705.124671] env[63372]: DEBUG nova.compute.manager [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 705.311130] env[63372]: DEBUG nova.network.neutron [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.454635] env[63372]: DEBUG nova.network.neutron [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Updating instance_info_cache with network_info: [{"id": "7688e644-e47b-4992-b037-343a71e18d2f", "address": "fa:16:3e:dc:d7:8d", "network": {"id": "d26c71aa-f3c4-4b1a-a263-fa95047e2b74", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1927982377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b68c21179fb24008baf1da0d0119fb29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6e2a9679-6746-40f2-951c-65fcd1af5f7b", "external-id": "nsx-vlan-transportzone-39", "segmentation_id": 39, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7688e644-e4", "ovs_interfaceid": "7688e644-e47b-4992-b037-343a71e18d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.488697] env[63372]: INFO nova.compute.manager [-] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Took 1.03 seconds to deallocate network for instance. [ 705.511929] env[63372]: INFO nova.compute.manager [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Took 34.97 seconds to build instance. [ 705.561136] env[63372]: DEBUG nova.scheduler.client.report [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.599638] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.957501] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Releasing lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.957822] env[63372]: DEBUG nova.compute.manager [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Instance network_info: |[{"id": "7688e644-e47b-4992-b037-343a71e18d2f", "address": "fa:16:3e:dc:d7:8d", "network": {"id": "d26c71aa-f3c4-4b1a-a263-fa95047e2b74", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1927982377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b68c21179fb24008baf1da0d0119fb29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6e2a9679-6746-40f2-951c-65fcd1af5f7b", "external-id": "nsx-vlan-transportzone-39", "segmentation_id": 39, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7688e644-e4", "ovs_interfaceid": "7688e644-e47b-4992-b037-343a71e18d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 705.958262] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:d7:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6e2a9679-6746-40f2-951c-65fcd1af5f7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7688e644-e47b-4992-b037-343a71e18d2f', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 705.966097] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Creating folder: Project (b68c21179fb24008baf1da0d0119fb29). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 705.966540] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-28579919-0c59-4319-9dd6-9d410365a132 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.979808] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Created folder: Project (b68c21179fb24008baf1da0d0119fb29) in parent group-v227230. [ 705.980089] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Creating folder: Instances. Parent ref: group-v227289. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 705.980289] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a4f51a3e-32ea-48b3-ae46-79e78ac96d9d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.991562] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Created folder: Instances in parent group-v227289. [ 705.991804] env[63372]: DEBUG oslo.service.loopingcall [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 705.991994] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 705.992218] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca031074-3383-4676-988d-091ec9e2ebb5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.007284] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.012493] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 706.012493] env[63372]: value = "task-1023835" [ 706.012493] env[63372]: _type = "Task" [ 706.012493] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.015823] env[63372]: DEBUG oslo_concurrency.lockutils [None req-86356b08-55d8-4bbf-9c67-3fe65ba8fcd0 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Lock "85883c62-6d5e-4fa6-b4f0-afac984dc6bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.713s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.025020] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023835, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.068063] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.952s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.070651] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.071s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.070883] env[63372]: DEBUG nova.objects.instance [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Lazy-loading 'resources' on Instance uuid 5adf6952-e019-4ea0-be91-8fe5fb83d53e {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 706.095898] env[63372]: INFO nova.scheduler.client.report [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Deleted allocations for instance 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60 [ 706.133765] env[63372]: DEBUG nova.compute.manager [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 706.159761] env[63372]: DEBUG nova.virt.hardware [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 706.160112] env[63372]: DEBUG nova.virt.hardware [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 706.160284] env[63372]: DEBUG nova.virt.hardware [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.160470] env[63372]: DEBUG nova.virt.hardware [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 706.160617] env[63372]: DEBUG nova.virt.hardware [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.160751] env[63372]: DEBUG nova.virt.hardware [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 706.160956] env[63372]: DEBUG nova.virt.hardware [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 706.161122] env[63372]: DEBUG nova.virt.hardware [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 706.161287] env[63372]: DEBUG nova.virt.hardware [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 706.161445] env[63372]: DEBUG nova.virt.hardware [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 706.161611] env[63372]: DEBUG nova.virt.hardware [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.162753] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77523d3f-1d9d-4238-aceb-6a9fceffc1c3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.172194] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de2a10f-f4bd-479d-b9af-38ef7028bbaa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.289783] env[63372]: DEBUG nova.compute.manager [req-c2804032-aea0-40da-84ae-995d10986b19 req-791686ef-9d83-40e1-b768-937acb2b3f60 service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Received event network-changed-7688e644-e47b-4992-b037-343a71e18d2f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 706.289934] env[63372]: DEBUG nova.compute.manager [req-c2804032-aea0-40da-84ae-995d10986b19 req-791686ef-9d83-40e1-b768-937acb2b3f60 service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Refreshing instance network info cache due to event network-changed-7688e644-e47b-4992-b037-343a71e18d2f. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 706.290216] env[63372]: DEBUG oslo_concurrency.lockutils [req-c2804032-aea0-40da-84ae-995d10986b19 req-791686ef-9d83-40e1-b768-937acb2b3f60 service nova] Acquiring lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.290302] env[63372]: DEBUG oslo_concurrency.lockutils [req-c2804032-aea0-40da-84ae-995d10986b19 req-791686ef-9d83-40e1-b768-937acb2b3f60 service nova] Acquired lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.290461] env[63372]: DEBUG nova.network.neutron [req-c2804032-aea0-40da-84ae-995d10986b19 req-791686ef-9d83-40e1-b768-937acb2b3f60 service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Refreshing network info cache for port 7688e644-e47b-4992-b037-343a71e18d2f {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 706.455508] env[63372]: DEBUG nova.network.neutron [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Successfully updated port: 620d14ec-96a7-4912-bb91-44d2c290d431 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 706.518214] env[63372]: DEBUG nova.compute.manager [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 706.523743] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023835, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.603835] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1dae3772-7971-4feb-8651-f2fb1d2b1c65 tempest-ImagesNegativeTestJSON-895649703 tempest-ImagesNegativeTestJSON-895649703-project-member] Lock "9404a0f0-a3e5-43ee-bc20-e9566cfe5a60" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.239s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.927264] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-353998f4-a5a1-4d52-8925-df5d6db51eea {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.935779] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0609dc-8672-477a-bc6c-e4b97e489e48 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.969292] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquiring lock "refresh_cache-c825d2e2-a9cc-4c7a-b92e-039756d4121d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.969449] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquired lock "refresh_cache-c825d2e2-a9cc-4c7a-b92e-039756d4121d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.969601] env[63372]: DEBUG nova.network.neutron [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.975945] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22eca237-a3aa-4a09-84b1-64c90febe232 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.985860] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ccbb5d-5b22-4eb3-8362-247ee8c14b44 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.001140] env[63372]: DEBUG nova.compute.provider_tree [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.026766] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023835, 'name': CreateVM_Task, 'duration_secs': 0.529589} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.026938] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 707.027629] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.027740] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.028524] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 707.028524] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-258c4f1d-7691-4611-8d24-9ceb2b93f5c7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.036090] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Waiting for the task: (returnval){ [ 707.036090] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5227eaf4-c214-40aa-a6b4-0b988e002fc9" [ 707.036090] env[63372]: _type = "Task" [ 707.036090] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.044794] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5227eaf4-c214-40aa-a6b4-0b988e002fc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.045690] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.193112] env[63372]: DEBUG nova.network.neutron [req-c2804032-aea0-40da-84ae-995d10986b19 req-791686ef-9d83-40e1-b768-937acb2b3f60 service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Updated VIF entry in instance network info cache for port 7688e644-e47b-4992-b037-343a71e18d2f. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 707.193112] env[63372]: DEBUG nova.network.neutron [req-c2804032-aea0-40da-84ae-995d10986b19 req-791686ef-9d83-40e1-b768-937acb2b3f60 service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Updating instance_info_cache with network_info: [{"id": "7688e644-e47b-4992-b037-343a71e18d2f", "address": "fa:16:3e:dc:d7:8d", "network": {"id": "d26c71aa-f3c4-4b1a-a263-fa95047e2b74", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1927982377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b68c21179fb24008baf1da0d0119fb29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6e2a9679-6746-40f2-951c-65fcd1af5f7b", "external-id": "nsx-vlan-transportzone-39", "segmentation_id": 39, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7688e644-e4", "ovs_interfaceid": "7688e644-e47b-4992-b037-343a71e18d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.200999] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Acquiring lock "85883c62-6d5e-4fa6-b4f0-afac984dc6bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.204730] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Lock "85883c62-6d5e-4fa6-b4f0-afac984dc6bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.204730] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Acquiring lock "85883c62-6d5e-4fa6-b4f0-afac984dc6bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.204730] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Lock "85883c62-6d5e-4fa6-b4f0-afac984dc6bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.204730] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Lock "85883c62-6d5e-4fa6-b4f0-afac984dc6bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.206061] env[63372]: INFO nova.compute.manager [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Terminating instance [ 707.209026] env[63372]: DEBUG nova.compute.manager [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 707.209026] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 707.209635] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735dd67e-94fd-4568-8b3d-a2eeab62aa0a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.219152] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 707.219442] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bfbb1ebc-9898-4ce5-9219-edf65bc073cb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.226935] env[63372]: DEBUG oslo_vmware.api [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Waiting for the task: (returnval){ [ 707.226935] env[63372]: value = "task-1023836" [ 707.226935] env[63372]: _type = "Task" [ 707.226935] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.235837] env[63372]: DEBUG oslo_vmware.api [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023836, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.504297] env[63372]: DEBUG nova.scheduler.client.report [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.513247] env[63372]: DEBUG nova.network.neutron [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.547757] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5227eaf4-c214-40aa-a6b4-0b988e002fc9, 'name': SearchDatastore_Task, 'duration_secs': 0.036423} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.548054] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.548282] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 707.548506] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.548647] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.548821] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 707.549087] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97b3e0ee-ed11-40a4-835a-57989cd77adf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.558637] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 707.559613] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 707.559613] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bd20996-1d5b-4d3e-a68b-e038c5effcf6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.568209] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Waiting for the task: (returnval){ [ 707.568209] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52062836-f7fd-316c-27b4-85bae4da8019" [ 707.568209] env[63372]: _type = "Task" [ 707.568209] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.574195] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52062836-f7fd-316c-27b4-85bae4da8019, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.694200] env[63372]: DEBUG oslo_concurrency.lockutils [req-c2804032-aea0-40da-84ae-995d10986b19 req-791686ef-9d83-40e1-b768-937acb2b3f60 service nova] Releasing lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.739317] env[63372]: DEBUG oslo_vmware.api [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023836, 'name': PowerOffVM_Task, 'duration_secs': 0.195818} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.740069] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 707.740300] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 707.740594] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4fcc5acb-1e4d-4dcc-94a2-844ed77b6e15 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.748323] env[63372]: DEBUG nova.network.neutron [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Updating instance_info_cache with network_info: [{"id": "620d14ec-96a7-4912-bb91-44d2c290d431", "address": "fa:16:3e:f4:07:ec", "network": {"id": "e5ec0f30-c191-4392-95c0-ac48df5cc431", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-828375451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b971b7d7a62d4902a930b10d96cfb51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9a1e09ef-7c9c-45d9-9bf4-55b913524948", "external-id": "nsx-vlan-transportzone-466", "segmentation_id": 466, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap620d14ec-96", "ovs_interfaceid": "620d14ec-96a7-4912-bb91-44d2c290d431", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.806334] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 707.806674] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 707.806894] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Deleting the datastore file [datastore2] 85883c62-6d5e-4fa6-b4f0-afac984dc6bb {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 707.807360] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0fc8d7ec-e3bb-4d51-afe2-f601061b7fa3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.815528] env[63372]: DEBUG oslo_vmware.api [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Waiting for the task: (returnval){ [ 707.815528] env[63372]: value = "task-1023838" [ 707.815528] env[63372]: _type = "Task" [ 707.815528] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.824150] env[63372]: DEBUG oslo_vmware.api [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023838, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.009479] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.939s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.012169] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.098s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.013691] env[63372]: INFO nova.compute.claims [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.034794] env[63372]: INFO nova.scheduler.client.report [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Deleted allocations for instance 5adf6952-e019-4ea0-be91-8fe5fb83d53e [ 708.078464] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52062836-f7fd-316c-27b4-85bae4da8019, 'name': SearchDatastore_Task, 'duration_secs': 0.010138} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.079185] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea5bdd4d-df02-4d69-93e6-c58cf21f0b5c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.086982] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Waiting for the task: (returnval){ [ 708.086982] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523c2495-40f0-ca3f-6e36-21436ff1b7f6" [ 708.086982] env[63372]: _type = "Task" [ 708.086982] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.095089] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523c2495-40f0-ca3f-6e36-21436ff1b7f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.250596] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Releasing lock "refresh_cache-c825d2e2-a9cc-4c7a-b92e-039756d4121d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.250934] env[63372]: DEBUG nova.compute.manager [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Instance network_info: |[{"id": "620d14ec-96a7-4912-bb91-44d2c290d431", "address": "fa:16:3e:f4:07:ec", "network": {"id": "e5ec0f30-c191-4392-95c0-ac48df5cc431", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-828375451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b971b7d7a62d4902a930b10d96cfb51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9a1e09ef-7c9c-45d9-9bf4-55b913524948", "external-id": "nsx-vlan-transportzone-466", "segmentation_id": 466, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap620d14ec-96", "ovs_interfaceid": "620d14ec-96a7-4912-bb91-44d2c290d431", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 708.251471] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f4:07:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9a1e09ef-7c9c-45d9-9bf4-55b913524948', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '620d14ec-96a7-4912-bb91-44d2c290d431', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 708.259159] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Creating folder: Project (b971b7d7a62d4902a930b10d96cfb51e). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 708.259509] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1f149da9-3ed1-4684-b7b9-8bc23c2d4608 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.272054] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Created folder: Project (b971b7d7a62d4902a930b10d96cfb51e) in parent group-v227230. [ 708.272262] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Creating folder: Instances. Parent ref: group-v227292. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 708.272612] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-16931b45-605d-4105-a408-deabdd7f4d17 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.282749] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Created folder: Instances in parent group-v227292. [ 708.282980] env[63372]: DEBUG oslo.service.loopingcall [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.283188] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 708.283502] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7f6c22f-02de-4c58-8c3c-ed850c71d248 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.305332] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 708.305332] env[63372]: value = "task-1023841" [ 708.305332] env[63372]: _type = "Task" [ 708.305332] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.314198] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023841, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.325017] env[63372]: DEBUG oslo_vmware.api [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Task: {'id': task-1023838, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156423} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.325316] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 708.325504] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 708.325678] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 708.325848] env[63372]: INFO nova.compute.manager [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Took 1.12 seconds to destroy the instance on the hypervisor. [ 708.326133] env[63372]: DEBUG oslo.service.loopingcall [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.326327] env[63372]: DEBUG nova.compute.manager [-] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.326424] env[63372]: DEBUG nova.network.neutron [-] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 708.425412] env[63372]: DEBUG nova.compute.manager [req-889f7819-aceb-48c3-8e4c-58169fb6570f req-7b215a4f-c446-439a-bbc7-aba8e5670a6b service nova] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Received event network-vif-plugged-620d14ec-96a7-4912-bb91-44d2c290d431 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.425641] env[63372]: DEBUG oslo_concurrency.lockutils [req-889f7819-aceb-48c3-8e4c-58169fb6570f req-7b215a4f-c446-439a-bbc7-aba8e5670a6b service nova] Acquiring lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.425937] env[63372]: DEBUG oslo_concurrency.lockutils [req-889f7819-aceb-48c3-8e4c-58169fb6570f req-7b215a4f-c446-439a-bbc7-aba8e5670a6b service nova] Lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.427029] env[63372]: DEBUG oslo_concurrency.lockutils [req-889f7819-aceb-48c3-8e4c-58169fb6570f req-7b215a4f-c446-439a-bbc7-aba8e5670a6b service nova] Lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.427029] env[63372]: DEBUG nova.compute.manager [req-889f7819-aceb-48c3-8e4c-58169fb6570f req-7b215a4f-c446-439a-bbc7-aba8e5670a6b service nova] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] No waiting events found dispatching network-vif-plugged-620d14ec-96a7-4912-bb91-44d2c290d431 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 708.427029] env[63372]: WARNING nova.compute.manager [req-889f7819-aceb-48c3-8e4c-58169fb6570f req-7b215a4f-c446-439a-bbc7-aba8e5670a6b service nova] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Received unexpected event network-vif-plugged-620d14ec-96a7-4912-bb91-44d2c290d431 for instance with vm_state building and task_state spawning. [ 708.427029] env[63372]: DEBUG nova.compute.manager [req-889f7819-aceb-48c3-8e4c-58169fb6570f req-7b215a4f-c446-439a-bbc7-aba8e5670a6b service nova] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Received event network-changed-620d14ec-96a7-4912-bb91-44d2c290d431 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.427029] env[63372]: DEBUG nova.compute.manager [req-889f7819-aceb-48c3-8e4c-58169fb6570f req-7b215a4f-c446-439a-bbc7-aba8e5670a6b service nova] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Refreshing instance network info cache due to event network-changed-620d14ec-96a7-4912-bb91-44d2c290d431. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 708.427541] env[63372]: DEBUG oslo_concurrency.lockutils [req-889f7819-aceb-48c3-8e4c-58169fb6570f req-7b215a4f-c446-439a-bbc7-aba8e5670a6b service nova] Acquiring lock "refresh_cache-c825d2e2-a9cc-4c7a-b92e-039756d4121d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.427541] env[63372]: DEBUG oslo_concurrency.lockutils [req-889f7819-aceb-48c3-8e4c-58169fb6570f req-7b215a4f-c446-439a-bbc7-aba8e5670a6b service nova] Acquired lock "refresh_cache-c825d2e2-a9cc-4c7a-b92e-039756d4121d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.427541] env[63372]: DEBUG nova.network.neutron [req-889f7819-aceb-48c3-8e4c-58169fb6570f req-7b215a4f-c446-439a-bbc7-aba8e5670a6b service nova] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Refreshing network info cache for port 620d14ec-96a7-4912-bb91-44d2c290d431 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 708.549307] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4339aea9-945e-4400-8a07-6fd100503548 tempest-InstanceActionsV221TestJSON-112867885 tempest-InstanceActionsV221TestJSON-112867885-project-member] Lock "5adf6952-e019-4ea0-be91-8fe5fb83d53e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.490s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.607539] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523c2495-40f0-ca3f-6e36-21436ff1b7f6, 'name': SearchDatastore_Task, 'duration_secs': 0.010992} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.608682] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.609169] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] e5522005-8c54-43e1-ae23-5e5ff1ef0ee9/e5522005-8c54-43e1-ae23-5e5ff1ef0ee9.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 708.609903] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49f8cb00-170e-4719-ad0e-b4c41f4cf611 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.619684] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Waiting for the task: (returnval){ [ 708.619684] env[63372]: value = "task-1023842" [ 708.619684] env[63372]: _type = "Task" [ 708.619684] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.629791] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023842, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.817672] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023841, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.110651] env[63372]: DEBUG nova.network.neutron [-] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.141427] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023842, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.288326] env[63372]: DEBUG nova.network.neutron [req-889f7819-aceb-48c3-8e4c-58169fb6570f req-7b215a4f-c446-439a-bbc7-aba8e5670a6b service nova] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Updated VIF entry in instance network info cache for port 620d14ec-96a7-4912-bb91-44d2c290d431. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 709.288799] env[63372]: DEBUG nova.network.neutron [req-889f7819-aceb-48c3-8e4c-58169fb6570f req-7b215a4f-c446-439a-bbc7-aba8e5670a6b service nova] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Updating instance_info_cache with network_info: [{"id": "620d14ec-96a7-4912-bb91-44d2c290d431", "address": "fa:16:3e:f4:07:ec", "network": {"id": "e5ec0f30-c191-4392-95c0-ac48df5cc431", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-828375451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b971b7d7a62d4902a930b10d96cfb51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9a1e09ef-7c9c-45d9-9bf4-55b913524948", "external-id": "nsx-vlan-transportzone-466", "segmentation_id": 466, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap620d14ec-96", "ovs_interfaceid": "620d14ec-96a7-4912-bb91-44d2c290d431", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.321588] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023841, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.445681] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ed6b85-c017-4049-8a40-637196fa6c66 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.454781] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adde8fe9-eb64-42e2-b78f-e98076d2416d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.486991] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a458e55f-fc2c-4c82-831e-87945b1e5b44 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.496891] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e2140e9-814f-4393-bbfe-b4b11ef28bb5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.512618] env[63372]: DEBUG nova.compute.provider_tree [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 709.614801] env[63372]: INFO nova.compute.manager [-] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Took 1.29 seconds to deallocate network for instance. [ 709.632729] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023842, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548954} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.633793] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] e5522005-8c54-43e1-ae23-5e5ff1ef0ee9/e5522005-8c54-43e1-ae23-5e5ff1ef0ee9.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 709.634037] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 709.635067] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5026f1f6-7fb1-434e-9ce1-18034d324a17 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.643141] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Waiting for the task: (returnval){ [ 709.643141] env[63372]: value = "task-1023843" [ 709.643141] env[63372]: _type = "Task" [ 709.643141] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.652329] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023843, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.792680] env[63372]: DEBUG oslo_concurrency.lockutils [req-889f7819-aceb-48c3-8e4c-58169fb6570f req-7b215a4f-c446-439a-bbc7-aba8e5670a6b service nova] Releasing lock "refresh_cache-c825d2e2-a9cc-4c7a-b92e-039756d4121d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.818549] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023841, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.033860] env[63372]: ERROR nova.scheduler.client.report [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [req-f72ba222-245f-4cb3-ba4f-2a2706c4bbfb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f72ba222-245f-4cb3-ba4f-2a2706c4bbfb"}]} [ 710.050505] env[63372]: DEBUG nova.scheduler.client.report [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Refreshing inventories for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 710.066106] env[63372]: DEBUG nova.scheduler.client.report [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Updating ProviderTree inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 710.066353] env[63372]: DEBUG nova.compute.provider_tree [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 710.080715] env[63372]: DEBUG nova.scheduler.client.report [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Refreshing aggregate associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, aggregates: None {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 710.103583] env[63372]: DEBUG nova.scheduler.client.report [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Refreshing trait associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 710.126545] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.155196] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023843, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078507} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.158789] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 710.160866] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120c1f2c-534e-44ae-8369-b0a3a19fbcc8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.191070] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] e5522005-8c54-43e1-ae23-5e5ff1ef0ee9/e5522005-8c54-43e1-ae23-5e5ff1ef0ee9.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 710.194210] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b481ba0-4244-4a86-a326-30d127864351 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.218407] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Waiting for the task: (returnval){ [ 710.218407] env[63372]: value = "task-1023844" [ 710.218407] env[63372]: _type = "Task" [ 710.218407] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.228180] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023844, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.321684] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023841, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.525156] env[63372]: DEBUG nova.compute.manager [req-739db771-0cb9-41f7-945d-c3b2df0c55de req-b7302adb-75a9-4624-86f5-21810dff553d service nova] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Received event network-vif-deleted-f47ddf2b-5ce1-46cc-aeac-18507b30f21e {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.553199] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605b56d1-57e2-474c-8cd3-4757fc8266e8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.561692] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d5a4a9-bbfb-40d2-be83-1ceae29bf988 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.600345] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf4f6ac-8f5d-4d97-9af5-6cc6e390446e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.608353] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d4eb56b-1132-4b5b-a971-a782d0d74789 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.623088] env[63372]: DEBUG nova.compute.provider_tree [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 710.730280] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023844, 'name': ReconfigVM_Task, 'duration_secs': 0.290521} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.730606] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Reconfigured VM instance instance-00000029 to attach disk [datastore1] e5522005-8c54-43e1-ae23-5e5ff1ef0ee9/e5522005-8c54-43e1-ae23-5e5ff1ef0ee9.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 710.731263] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-945db4ba-aefc-4558-b8b8-1e57286798a3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.739296] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Waiting for the task: (returnval){ [ 710.739296] env[63372]: value = "task-1023845" [ 710.739296] env[63372]: _type = "Task" [ 710.739296] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.751951] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023845, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.819443] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023841, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.162594] env[63372]: DEBUG nova.scheduler.client.report [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 69 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 711.163230] env[63372]: DEBUG nova.compute.provider_tree [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 69 to 70 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 711.163230] env[63372]: DEBUG nova.compute.provider_tree [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 711.252332] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023845, 'name': Rename_Task, 'duration_secs': 0.151678} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.252458] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 711.252697] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66be1316-a2a5-4d11-88af-eed3eff19cf2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.263589] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Waiting for the task: (returnval){ [ 711.263589] env[63372]: value = "task-1023846" [ 711.263589] env[63372]: _type = "Task" [ 711.263589] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.272522] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023846, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.320045] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023841, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.670278] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.658s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.670688] env[63372]: DEBUG nova.compute.manager [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 711.674405] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.287s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.676182] env[63372]: INFO nova.compute.claims [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 711.774216] env[63372]: DEBUG oslo_vmware.api [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023846, 'name': PowerOnVM_Task, 'duration_secs': 0.501009} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.774429] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 711.774627] env[63372]: INFO nova.compute.manager [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Took 8.23 seconds to spawn the instance on the hypervisor. [ 711.774790] env[63372]: DEBUG nova.compute.manager [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 711.775577] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e92dc441-50bb-43c4-86c1-db6d020a172e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.821705] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023841, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.191575] env[63372]: DEBUG nova.compute.utils [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 712.197580] env[63372]: DEBUG nova.compute.manager [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 712.198093] env[63372]: DEBUG nova.network.neutron [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 712.273996] env[63372]: DEBUG nova.policy [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3fb6017431c4a16b1599dae8df84311', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6dd3c2f3447d41fe8434faa68ad5ea0c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 712.298705] env[63372]: INFO nova.compute.manager [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Took 41.19 seconds to build instance. [ 712.324835] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023841, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.704826] env[63372]: DEBUG nova.network.neutron [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Successfully created port: 62ba51bf-2e6c-4205-93c0-e45bfd98f6ae {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 712.707103] env[63372]: DEBUG nova.compute.manager [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 712.800823] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd66bab0-72ae-41ca-a08d-eb76b8c9b8cd tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.538s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.826010] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023841, 'name': CreateVM_Task, 'duration_secs': 4.094106} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.826315] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 712.827850] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.827850] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.827850] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 712.827850] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7231083b-3533-43b6-80f1-8451a35552fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.832863] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Waiting for the task: (returnval){ [ 712.832863] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525133ed-2898-b6d3-6858-4660187fbc4a" [ 712.832863] env[63372]: _type = "Task" [ 712.832863] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.844726] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525133ed-2898-b6d3-6858-4660187fbc4a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.118964] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7810e9-8fd4-4855-b3cb-cf3cab0e2af6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.130894] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4081a6b3-b719-4f83-a3e4-e1dc76fd600b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.172181] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789add26-5612-4c50-a034-3abc0b87fa85 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.181198] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfa3797-2657-45af-b7ac-1bffbbeb53df {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.198308] env[63372]: DEBUG nova.compute.provider_tree [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.212273] env[63372]: INFO nova.virt.block_device [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Booting with volume a483fa3d-7a97-48fa-8fe5-723336b546a9 at /dev/sda [ 713.265689] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0d8bfdc8-2a28-4144-9246-54d7662b93a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.275939] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f62a141-b2ba-4cf0-8ed5-e0c8c6eedeb4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.306189] env[63372]: DEBUG nova.compute.manager [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 713.308831] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ba4e4dd1-47b6-490d-a5ef-45eb31ef0c17 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.318258] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d79593-094c-4310-80ff-01fcff0b0a41 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.344204] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525133ed-2898-b6d3-6858-4660187fbc4a, 'name': SearchDatastore_Task, 'duration_secs': 0.011025} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.353078] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.353347] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 713.353581] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.355149] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.355149] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 713.355149] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6d34cdd1-2cc2-455f-8792-c4fbdc821cd8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.357009] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22b466e-95bc-47e5-b466-bf228099889c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.365248] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e214338-45b7-4489-b436-1f4ac6eabbe9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.369547] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 713.369730] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 713.370825] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-054838f0-a441-450a-9031-32c4d324b32c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.384034] env[63372]: DEBUG nova.virt.block_device [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Updating existing volume attachment record: 785db812-0c23-491d-a80e-a19e80cf9f6c {{(pid=63372) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 713.386711] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Waiting for the task: (returnval){ [ 713.386711] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5231075c-047b-4ae1-e382-6f1e0dfb9865" [ 713.386711] env[63372]: _type = "Task" [ 713.386711] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.395590] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5231075c-047b-4ae1-e382-6f1e0dfb9865, 'name': SearchDatastore_Task, 'duration_secs': 0.01439} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.396514] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6f576fe-6e31-457a-a74b-55b1802a0f42 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.402266] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Waiting for the task: (returnval){ [ 713.402266] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5265ffc5-bdab-8fe7-a1ef-2980baef891c" [ 713.402266] env[63372]: _type = "Task" [ 713.402266] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.410896] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5265ffc5-bdab-8fe7-a1ef-2980baef891c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.677293] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Acquiring lock "4c9eb955-3bed-4b26-866f-b2a876b835dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.677448] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Lock "4c9eb955-3bed-4b26-866f-b2a876b835dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.701200] env[63372]: DEBUG nova.scheduler.client.report [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 713.839317] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.914978] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5265ffc5-bdab-8fe7-a1ef-2980baef891c, 'name': SearchDatastore_Task, 'duration_secs': 0.010389} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.915395] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.915661] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] c825d2e2-a9cc-4c7a-b92e-039756d4121d/c825d2e2-a9cc-4c7a-b92e-039756d4121d.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 713.915917] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-66a62ef4-9f6d-4fb5-add9-732f516cb6c9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.923828] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Waiting for the task: (returnval){ [ 713.923828] env[63372]: value = "task-1023847" [ 713.923828] env[63372]: _type = "Task" [ 713.923828] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.934675] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023847, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.209231] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.211404] env[63372]: DEBUG nova.compute.manager [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 714.215111] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.646s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.217221] env[63372]: INFO nova.compute.claims [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.263296] env[63372]: DEBUG nova.compute.manager [req-abd5cab2-b741-4fd6-841a-9af9b5af5cdd req-b78defbe-0b9e-4a5b-9d56-fbfd05a4de8d service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Received event network-changed-7688e644-e47b-4992-b037-343a71e18d2f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.263463] env[63372]: DEBUG nova.compute.manager [req-abd5cab2-b741-4fd6-841a-9af9b5af5cdd req-b78defbe-0b9e-4a5b-9d56-fbfd05a4de8d service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Refreshing instance network info cache due to event network-changed-7688e644-e47b-4992-b037-343a71e18d2f. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 714.263607] env[63372]: DEBUG oslo_concurrency.lockutils [req-abd5cab2-b741-4fd6-841a-9af9b5af5cdd req-b78defbe-0b9e-4a5b-9d56-fbfd05a4de8d service nova] Acquiring lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.265360] env[63372]: DEBUG oslo_concurrency.lockutils [req-abd5cab2-b741-4fd6-841a-9af9b5af5cdd req-b78defbe-0b9e-4a5b-9d56-fbfd05a4de8d service nova] Acquired lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.265789] env[63372]: DEBUG nova.network.neutron [req-abd5cab2-b741-4fd6-841a-9af9b5af5cdd req-b78defbe-0b9e-4a5b-9d56-fbfd05a4de8d service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Refreshing network info cache for port 7688e644-e47b-4992-b037-343a71e18d2f {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 714.320287] env[63372]: DEBUG nova.network.neutron [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Successfully updated port: 62ba51bf-2e6c-4205-93c0-e45bfd98f6ae {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 714.361714] env[63372]: DEBUG nova.compute.manager [req-f087942d-43f5-4559-8118-b4ea344ae3ea req-303b15c4-c18a-4d06-9053-8776376852bb service nova] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Received event network-vif-plugged-62ba51bf-2e6c-4205-93c0-e45bfd98f6ae {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.361939] env[63372]: DEBUG oslo_concurrency.lockutils [req-f087942d-43f5-4559-8118-b4ea344ae3ea req-303b15c4-c18a-4d06-9053-8776376852bb service nova] Acquiring lock "5d819bd2-f7f9-480d-83ae-fbd4d2e0d382-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.362578] env[63372]: DEBUG oslo_concurrency.lockutils [req-f087942d-43f5-4559-8118-b4ea344ae3ea req-303b15c4-c18a-4d06-9053-8776376852bb service nova] Lock "5d819bd2-f7f9-480d-83ae-fbd4d2e0d382-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.362578] env[63372]: DEBUG oslo_concurrency.lockutils [req-f087942d-43f5-4559-8118-b4ea344ae3ea req-303b15c4-c18a-4d06-9053-8776376852bb service nova] Lock "5d819bd2-f7f9-480d-83ae-fbd4d2e0d382-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.362578] env[63372]: DEBUG nova.compute.manager [req-f087942d-43f5-4559-8118-b4ea344ae3ea req-303b15c4-c18a-4d06-9053-8776376852bb service nova] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] No waiting events found dispatching network-vif-plugged-62ba51bf-2e6c-4205-93c0-e45bfd98f6ae {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 714.363636] env[63372]: WARNING nova.compute.manager [req-f087942d-43f5-4559-8118-b4ea344ae3ea req-303b15c4-c18a-4d06-9053-8776376852bb service nova] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Received unexpected event network-vif-plugged-62ba51bf-2e6c-4205-93c0-e45bfd98f6ae for instance with vm_state building and task_state block_device_mapping. [ 714.435699] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023847, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478655} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.435699] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] c825d2e2-a9cc-4c7a-b92e-039756d4121d/c825d2e2-a9cc-4c7a-b92e-039756d4121d.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 714.436073] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 714.436176] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c75d60bd-6ff3-421f-9c6d-a627a9bb53fc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.443687] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Waiting for the task: (returnval){ [ 714.443687] env[63372]: value = "task-1023848" [ 714.443687] env[63372]: _type = "Task" [ 714.443687] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.455864] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023848, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.724709] env[63372]: DEBUG nova.compute.utils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 714.728020] env[63372]: DEBUG nova.compute.manager [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 714.728020] env[63372]: DEBUG nova.network.neutron [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 714.776527] env[63372]: DEBUG nova.policy [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '881df5e623b340a487f37076c0cd4a15', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b6afb5579f2b48e4b094f6490136417f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 714.824122] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Acquiring lock "refresh_cache-5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.824122] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Acquired lock "refresh_cache-5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.824122] env[63372]: DEBUG nova.network.neutron [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 714.957850] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023848, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065432} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.961279] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 714.962420] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a5e480-1d04-4d05-9510-67d2fcc4bcf4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.991610] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] c825d2e2-a9cc-4c7a-b92e-039756d4121d/c825d2e2-a9cc-4c7a-b92e-039756d4121d.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 714.992797] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5659bb5-5beb-4c8a-aca9-4efa11c3053f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.016787] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.017038] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.017276] env[63372]: INFO nova.compute.manager [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Rebooting instance [ 715.029408] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Waiting for the task: (returnval){ [ 715.029408] env[63372]: value = "task-1023849" [ 715.029408] env[63372]: _type = "Task" [ 715.029408] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.044530] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023849, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.152158] env[63372]: DEBUG nova.network.neutron [req-abd5cab2-b741-4fd6-841a-9af9b5af5cdd req-b78defbe-0b9e-4a5b-9d56-fbfd05a4de8d service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Updated VIF entry in instance network info cache for port 7688e644-e47b-4992-b037-343a71e18d2f. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 715.152525] env[63372]: DEBUG nova.network.neutron [req-abd5cab2-b741-4fd6-841a-9af9b5af5cdd req-b78defbe-0b9e-4a5b-9d56-fbfd05a4de8d service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Updating instance_info_cache with network_info: [{"id": "7688e644-e47b-4992-b037-343a71e18d2f", "address": "fa:16:3e:dc:d7:8d", "network": {"id": "d26c71aa-f3c4-4b1a-a263-fa95047e2b74", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1927982377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b68c21179fb24008baf1da0d0119fb29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6e2a9679-6746-40f2-951c-65fcd1af5f7b", "external-id": "nsx-vlan-transportzone-39", "segmentation_id": 39, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7688e644-e4", "ovs_interfaceid": "7688e644-e47b-4992-b037-343a71e18d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.217345] env[63372]: DEBUG nova.network.neutron [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Successfully created port: 022ca5fc-1544-4248-bb33-e377ab46b598 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.232746] env[63372]: DEBUG nova.compute.manager [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 715.368462] env[63372]: DEBUG nova.network.neutron [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.527171] env[63372]: DEBUG nova.compute.manager [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 715.528163] env[63372]: DEBUG nova.virt.hardware [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 715.528554] env[63372]: DEBUG nova.virt.hardware [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 715.528812] env[63372]: DEBUG nova.virt.hardware [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 715.531070] env[63372]: DEBUG nova.virt.hardware [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 715.531070] env[63372]: DEBUG nova.virt.hardware [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 715.531070] env[63372]: DEBUG nova.virt.hardware [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 715.531070] env[63372]: DEBUG nova.virt.hardware [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 715.531070] env[63372]: DEBUG nova.virt.hardware [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 715.531070] env[63372]: DEBUG nova.virt.hardware [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 715.531070] env[63372]: DEBUG nova.virt.hardware [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 715.531070] env[63372]: DEBUG nova.virt.hardware [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 715.531349] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e02b584-9ba7-4ecc-a547-b5d4a76339c3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.549087] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023849, 'name': ReconfigVM_Task, 'duration_secs': 0.300959} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.552574] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.552880] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Reconfigured VM instance instance-0000002a to attach disk [datastore1] c825d2e2-a9cc-4c7a-b92e-039756d4121d/c825d2e2-a9cc-4c7a-b92e-039756d4121d.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 715.554645] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-590d5a5a-39a5-4548-9ede-45261d900c40 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.559997] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1f04f372-f35d-4af6-bcfa-8242073ca2a9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.576761] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Waiting for the task: (returnval){ [ 715.576761] env[63372]: value = "task-1023850" [ 715.576761] env[63372]: _type = "Task" [ 715.576761] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.591754] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023850, 'name': Rename_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.614282] env[63372]: DEBUG nova.network.neutron [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Updating instance_info_cache with network_info: [{"id": "62ba51bf-2e6c-4205-93c0-e45bfd98f6ae", "address": "fa:16:3e:ea:91:de", "network": {"id": "3f4f7142-9adb-40bd-a4c4-50da438a2a82", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-657036764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6dd3c2f3447d41fe8434faa68ad5ea0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62ba51bf-2e", "ovs_interfaceid": "62ba51bf-2e6c-4205-93c0-e45bfd98f6ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.622164] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "bad91d77-c7ee-4572-b1ed-068b2a55233c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.622406] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "bad91d77-c7ee-4572-b1ed-068b2a55233c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.655686] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "cbd55518-a3b2-4636-ba43-c279fdd0bf8d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.656070] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "cbd55518-a3b2-4636-ba43-c279fdd0bf8d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.657515] env[63372]: DEBUG oslo_concurrency.lockutils [req-abd5cab2-b741-4fd6-841a-9af9b5af5cdd req-b78defbe-0b9e-4a5b-9d56-fbfd05a4de8d service nova] Releasing lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.658067] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquired lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.658260] env[63372]: DEBUG nova.network.neutron [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 715.690314] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57dbb0a9-3105-4e64-bb70-3147f6442cdd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.698782] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f190f7af-e823-4914-ba44-9c70c75e7eba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.736019] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c90751-0ea1-4c29-a23a-c0b7f0deb058 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.751182] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcca08fb-04e5-428b-be5f-bd80c1fed5ef {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.768411] env[63372]: DEBUG nova.compute.provider_tree [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.087640] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023850, 'name': Rename_Task, 'duration_secs': 0.147888} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.087930] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 716.088175] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e64dd872-3298-4188-be3d-07ce6a6235e4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.095693] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Waiting for the task: (returnval){ [ 716.095693] env[63372]: value = "task-1023851" [ 716.095693] env[63372]: _type = "Task" [ 716.095693] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.103736] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023851, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.120524] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Releasing lock "refresh_cache-5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.120872] env[63372]: DEBUG nova.compute.manager [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Instance network_info: |[{"id": "62ba51bf-2e6c-4205-93c0-e45bfd98f6ae", "address": "fa:16:3e:ea:91:de", "network": {"id": "3f4f7142-9adb-40bd-a4c4-50da438a2a82", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-657036764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6dd3c2f3447d41fe8434faa68ad5ea0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62ba51bf-2e", "ovs_interfaceid": "62ba51bf-2e6c-4205-93c0-e45bfd98f6ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 716.121306] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:91:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '62ba51bf-2e6c-4205-93c0-e45bfd98f6ae', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 716.129022] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Creating folder: Project (6dd3c2f3447d41fe8434faa68ad5ea0c). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 716.129331] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7021573e-d375-47c9-8e48-25b957b73b2d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.142913] env[63372]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 716.143100] env[63372]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63372) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 716.143447] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Folder already exists: Project (6dd3c2f3447d41fe8434faa68ad5ea0c). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 716.143635] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Creating folder: Instances. Parent ref: group-v227231. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 716.143861] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5dadbe9b-79f7-4d7a-9705-98d0ab79d081 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.153823] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Created folder: Instances in parent group-v227231. [ 716.154082] env[63372]: DEBUG oslo.service.loopingcall [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.154317] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 716.154532] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a224bd6b-8556-428d-bdca-6c9312a1883b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.177579] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 716.177579] env[63372]: value = "task-1023854" [ 716.177579] env[63372]: _type = "Task" [ 716.177579] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.186887] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023854, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.243317] env[63372]: DEBUG nova.compute.manager [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 716.270923] env[63372]: DEBUG nova.virt.hardware [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.271186] env[63372]: DEBUG nova.virt.hardware [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.271369] env[63372]: DEBUG nova.virt.hardware [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.271571] env[63372]: DEBUG nova.virt.hardware [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.271715] env[63372]: DEBUG nova.virt.hardware [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.271900] env[63372]: DEBUG nova.virt.hardware [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.272150] env[63372]: DEBUG nova.virt.hardware [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.272342] env[63372]: DEBUG nova.virt.hardware [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.272550] env[63372]: DEBUG nova.virt.hardware [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.272744] env[63372]: DEBUG nova.virt.hardware [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.272955] env[63372]: DEBUG nova.virt.hardware [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.273807] env[63372]: DEBUG nova.scheduler.client.report [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.278317] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fff369a-d067-46a2-99af-4a5c2dcdd155 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.287076] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35321d11-c761-402e-a592-8b6b29470e92 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.400845] env[63372]: DEBUG nova.network.neutron [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Updating instance_info_cache with network_info: [{"id": "7688e644-e47b-4992-b037-343a71e18d2f", "address": "fa:16:3e:dc:d7:8d", "network": {"id": "d26c71aa-f3c4-4b1a-a263-fa95047e2b74", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1927982377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b68c21179fb24008baf1da0d0119fb29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6e2a9679-6746-40f2-951c-65fcd1af5f7b", "external-id": "nsx-vlan-transportzone-39", "segmentation_id": 39, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7688e644-e4", "ovs_interfaceid": "7688e644-e47b-4992-b037-343a71e18d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.439390] env[63372]: DEBUG nova.compute.manager [req-4b746b4d-3252-4bc1-bb7f-9c80ced1608a req-a8804654-ad0a-422d-ad7f-7634fbb0f654 service nova] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Received event network-changed-62ba51bf-2e6c-4205-93c0-e45bfd98f6ae {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 716.439636] env[63372]: DEBUG nova.compute.manager [req-4b746b4d-3252-4bc1-bb7f-9c80ced1608a req-a8804654-ad0a-422d-ad7f-7634fbb0f654 service nova] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Refreshing instance network info cache due to event network-changed-62ba51bf-2e6c-4205-93c0-e45bfd98f6ae. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 716.439835] env[63372]: DEBUG oslo_concurrency.lockutils [req-4b746b4d-3252-4bc1-bb7f-9c80ced1608a req-a8804654-ad0a-422d-ad7f-7634fbb0f654 service nova] Acquiring lock "refresh_cache-5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.439978] env[63372]: DEBUG oslo_concurrency.lockutils [req-4b746b4d-3252-4bc1-bb7f-9c80ced1608a req-a8804654-ad0a-422d-ad7f-7634fbb0f654 service nova] Acquired lock "refresh_cache-5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.440236] env[63372]: DEBUG nova.network.neutron [req-4b746b4d-3252-4bc1-bb7f-9c80ced1608a req-a8804654-ad0a-422d-ad7f-7634fbb0f654 service nova] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Refreshing network info cache for port 62ba51bf-2e6c-4205-93c0-e45bfd98f6ae {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 716.609455] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023851, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.689355] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023854, 'name': CreateVM_Task, 'duration_secs': 0.366181} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.689355] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 716.689801] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': '785db812-0c23-491d-a80e-a19e80cf9f6c', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227234', 'volume_id': 'a483fa3d-7a97-48fa-8fe5-723336b546a9', 'name': 'volume-a483fa3d-7a97-48fa-8fe5-723336b546a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5d819bd2-f7f9-480d-83ae-fbd4d2e0d382', 'attached_at': '', 'detached_at': '', 'volume_id': 'a483fa3d-7a97-48fa-8fe5-723336b546a9', 'serial': 'a483fa3d-7a97-48fa-8fe5-723336b546a9'}, 'boot_index': 0, 'disk_bus': None, 'device_type': None, 'mount_device': '/dev/sda', 'guest_format': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=63372) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 716.690019] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Root volume attach. Driver type: vmdk {{(pid=63372) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 716.690846] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99856468-c146-43d8-9183-a550b016c9dd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.699661] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55727eed-f0d8-4136-a4cb-d8ed74f126c3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.706790] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30dccdbf-5387-4963-954a-dd5e24bd30be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.713983] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-f9edaf35-7aca-4890-892a-61a30ac1f47b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.721983] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Waiting for the task: (returnval){ [ 716.721983] env[63372]: value = "task-1023855" [ 716.721983] env[63372]: _type = "Task" [ 716.721983] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.730549] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023855, 'name': RelocateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.778759] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.564s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.779295] env[63372]: DEBUG nova.compute.manager [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 716.781821] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.113s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.783419] env[63372]: INFO nova.compute.claims [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 716.852936] env[63372]: DEBUG nova.network.neutron [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Successfully updated port: 022ca5fc-1544-4248-bb33-e377ab46b598 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 716.904398] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Releasing lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.906730] env[63372]: DEBUG nova.compute.manager [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 716.907748] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89e8319-295e-4b79-9c0c-52fecf9126f5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.112254] env[63372]: DEBUG oslo_vmware.api [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023851, 'name': PowerOnVM_Task, 'duration_secs': 0.52609} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.112627] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 717.112685] env[63372]: INFO nova.compute.manager [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Took 10.98 seconds to spawn the instance on the hypervisor. [ 717.112861] env[63372]: DEBUG nova.compute.manager [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 717.113800] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f493c9-b19f-444a-a4b1-bfb5258907b3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.171415] env[63372]: DEBUG nova.network.neutron [req-4b746b4d-3252-4bc1-bb7f-9c80ced1608a req-a8804654-ad0a-422d-ad7f-7634fbb0f654 service nova] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Updated VIF entry in instance network info cache for port 62ba51bf-2e6c-4205-93c0-e45bfd98f6ae. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 717.171775] env[63372]: DEBUG nova.network.neutron [req-4b746b4d-3252-4bc1-bb7f-9c80ced1608a req-a8804654-ad0a-422d-ad7f-7634fbb0f654 service nova] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Updating instance_info_cache with network_info: [{"id": "62ba51bf-2e6c-4205-93c0-e45bfd98f6ae", "address": "fa:16:3e:ea:91:de", "network": {"id": "3f4f7142-9adb-40bd-a4c4-50da438a2a82", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-657036764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6dd3c2f3447d41fe8434faa68ad5ea0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62ba51bf-2e", "ovs_interfaceid": "62ba51bf-2e6c-4205-93c0-e45bfd98f6ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.237592] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023855, 'name': RelocateVM_Task} progress is 42%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.287923] env[63372]: DEBUG nova.compute.utils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 717.291452] env[63372]: DEBUG nova.compute.manager [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 717.291660] env[63372]: DEBUG nova.network.neutron [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 717.357203] env[63372]: DEBUG nova.policy [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b96f55f1271346248fe83e2869aa90e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'de7a1fac8ec941b8b00357f62da80927', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 717.359298] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "refresh_cache-63b9095c-fb49-4dc1-a6dc-96529aeaab81" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.359298] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquired lock "refresh_cache-63b9095c-fb49-4dc1-a6dc-96529aeaab81" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.359475] env[63372]: DEBUG nova.network.neutron [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 717.641861] env[63372]: INFO nova.compute.manager [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Took 44.22 seconds to build instance. [ 717.675373] env[63372]: DEBUG oslo_concurrency.lockutils [req-4b746b4d-3252-4bc1-bb7f-9c80ced1608a req-a8804654-ad0a-422d-ad7f-7634fbb0f654 service nova] Releasing lock "refresh_cache-5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.731766] env[63372]: DEBUG nova.network.neutron [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Successfully created port: c56eef25-45d9-459d-b839-c069b18a613c {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.740195] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023855, 'name': RelocateVM_Task} progress is 54%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.795274] env[63372]: DEBUG nova.compute.manager [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 717.905581] env[63372]: DEBUG nova.network.neutron [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.926777] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e6c28f-f66a-4c89-b6fc-a00c30b8da2a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.945187] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Doing hard reboot of VM {{(pid=63372) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 717.945187] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-5bb2f44e-0e41-4302-bf37-241f3653f9fe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.958182] env[63372]: DEBUG oslo_vmware.api [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Waiting for the task: (returnval){ [ 717.958182] env[63372]: value = "task-1023856" [ 717.958182] env[63372]: _type = "Task" [ 717.958182] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.975371] env[63372]: DEBUG oslo_vmware.api [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023856, 'name': ResetVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.143907] env[63372]: DEBUG oslo_concurrency.lockutils [None req-32671f9b-b0bf-48eb-98ec-177e1cecd5b4 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.069s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.149557] env[63372]: DEBUG nova.network.neutron [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Updating instance_info_cache with network_info: [{"id": "022ca5fc-1544-4248-bb33-e377ab46b598", "address": "fa:16:3e:57:ea:33", "network": {"id": "b924481a-4e51-4483-a710-f01fc39c851c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1289474667-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b6afb5579f2b48e4b094f6490136417f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap022ca5fc-15", "ovs_interfaceid": "022ca5fc-1544-4248-bb33-e377ab46b598", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.246308] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023855, 'name': RelocateVM_Task} progress is 67%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.302545] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28fb13bb-d1df-4b34-8fd1-4a1072172a7f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.315562] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9dc137-a3bc-493b-8be8-bb2867c8e612 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.358433] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b0c5be-8e60-46b2-b203-5acfe2ab59c3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.371702] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd7e91b3-8ba7-496c-82d2-daade3ca79dc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.390868] env[63372]: DEBUG nova.compute.provider_tree [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.473083] env[63372]: DEBUG oslo_vmware.api [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023856, 'name': ResetVM_Task, 'duration_secs': 0.156692} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.473393] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Did hard reboot of VM {{(pid=63372) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 718.473566] env[63372]: DEBUG nova.compute.manager [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 718.474488] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f3cda19-ad06-4c3f-aa63-3ade40ccda84 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.650443] env[63372]: DEBUG nova.compute.manager [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 718.654030] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Releasing lock "refresh_cache-63b9095c-fb49-4dc1-a6dc-96529aeaab81" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.654460] env[63372]: DEBUG nova.compute.manager [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Instance network_info: |[{"id": "022ca5fc-1544-4248-bb33-e377ab46b598", "address": "fa:16:3e:57:ea:33", "network": {"id": "b924481a-4e51-4483-a710-f01fc39c851c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1289474667-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b6afb5579f2b48e4b094f6490136417f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap022ca5fc-15", "ovs_interfaceid": "022ca5fc-1544-4248-bb33-e377ab46b598", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 718.654848] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquiring lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.655082] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.655300] env[63372]: INFO nova.compute.manager [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Rebooting instance [ 718.658050] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:ea:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '40c947c4-f471-4d48-8e43-fee54198107e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '022ca5fc-1544-4248-bb33-e377ab46b598', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 718.666394] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Creating folder: Project (b6afb5579f2b48e4b094f6490136417f). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 718.668523] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bdc37005-69b8-4852-aa23-c328a22662e4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.672124] env[63372]: DEBUG nova.compute.manager [req-4d91feaa-bcf7-4c2d-8f29-0b3683a46711 req-86b78ffe-f5b7-48a7-a388-2a5a7ed126d9 service nova] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Received event network-vif-plugged-022ca5fc-1544-4248-bb33-e377ab46b598 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.672240] env[63372]: DEBUG oslo_concurrency.lockutils [req-4d91feaa-bcf7-4c2d-8f29-0b3683a46711 req-86b78ffe-f5b7-48a7-a388-2a5a7ed126d9 service nova] Acquiring lock "63b9095c-fb49-4dc1-a6dc-96529aeaab81-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.672380] env[63372]: DEBUG oslo_concurrency.lockutils [req-4d91feaa-bcf7-4c2d-8f29-0b3683a46711 req-86b78ffe-f5b7-48a7-a388-2a5a7ed126d9 service nova] Lock "63b9095c-fb49-4dc1-a6dc-96529aeaab81-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.672541] env[63372]: DEBUG oslo_concurrency.lockutils [req-4d91feaa-bcf7-4c2d-8f29-0b3683a46711 req-86b78ffe-f5b7-48a7-a388-2a5a7ed126d9 service nova] Lock "63b9095c-fb49-4dc1-a6dc-96529aeaab81-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.672700] env[63372]: DEBUG nova.compute.manager [req-4d91feaa-bcf7-4c2d-8f29-0b3683a46711 req-86b78ffe-f5b7-48a7-a388-2a5a7ed126d9 service nova] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] No waiting events found dispatching network-vif-plugged-022ca5fc-1544-4248-bb33-e377ab46b598 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 718.672856] env[63372]: WARNING nova.compute.manager [req-4d91feaa-bcf7-4c2d-8f29-0b3683a46711 req-86b78ffe-f5b7-48a7-a388-2a5a7ed126d9 service nova] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Received unexpected event network-vif-plugged-022ca5fc-1544-4248-bb33-e377ab46b598 for instance with vm_state building and task_state spawning. [ 718.673016] env[63372]: DEBUG nova.compute.manager [req-4d91feaa-bcf7-4c2d-8f29-0b3683a46711 req-86b78ffe-f5b7-48a7-a388-2a5a7ed126d9 service nova] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Received event network-changed-022ca5fc-1544-4248-bb33-e377ab46b598 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.673172] env[63372]: DEBUG nova.compute.manager [req-4d91feaa-bcf7-4c2d-8f29-0b3683a46711 req-86b78ffe-f5b7-48a7-a388-2a5a7ed126d9 service nova] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Refreshing instance network info cache due to event network-changed-022ca5fc-1544-4248-bb33-e377ab46b598. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 718.673501] env[63372]: DEBUG oslo_concurrency.lockutils [req-4d91feaa-bcf7-4c2d-8f29-0b3683a46711 req-86b78ffe-f5b7-48a7-a388-2a5a7ed126d9 service nova] Acquiring lock "refresh_cache-63b9095c-fb49-4dc1-a6dc-96529aeaab81" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.673640] env[63372]: DEBUG oslo_concurrency.lockutils [req-4d91feaa-bcf7-4c2d-8f29-0b3683a46711 req-86b78ffe-f5b7-48a7-a388-2a5a7ed126d9 service nova] Acquired lock "refresh_cache-63b9095c-fb49-4dc1-a6dc-96529aeaab81" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.673792] env[63372]: DEBUG nova.network.neutron [req-4d91feaa-bcf7-4c2d-8f29-0b3683a46711 req-86b78ffe-f5b7-48a7-a388-2a5a7ed126d9 service nova] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Refreshing network info cache for port 022ca5fc-1544-4248-bb33-e377ab46b598 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 718.690723] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Created folder: Project (b6afb5579f2b48e4b094f6490136417f) in parent group-v227230. [ 718.690931] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Creating folder: Instances. Parent ref: group-v227297. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 718.691487] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3b6bfc77-6046-4ed2-9e64-a8e3fdf5963a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.707435] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Created folder: Instances in parent group-v227297. [ 718.707695] env[63372]: DEBUG oslo.service.loopingcall [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.707959] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 718.708207] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-76f3141d-9f49-4f61-948c-92f103c2d32f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.735278] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 718.735278] env[63372]: value = "task-1023859" [ 718.735278] env[63372]: _type = "Task" [ 718.735278] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.745764] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023855, 'name': RelocateVM_Task} progress is 81%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.749048] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023859, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.807425] env[63372]: DEBUG nova.compute.manager [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 718.839447] env[63372]: DEBUG nova.virt.hardware [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 718.839750] env[63372]: DEBUG nova.virt.hardware [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 718.839971] env[63372]: DEBUG nova.virt.hardware [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.840172] env[63372]: DEBUG nova.virt.hardware [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 718.840353] env[63372]: DEBUG nova.virt.hardware [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.840530] env[63372]: DEBUG nova.virt.hardware [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 718.840800] env[63372]: DEBUG nova.virt.hardware [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 718.840934] env[63372]: DEBUG nova.virt.hardware [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 718.841333] env[63372]: DEBUG nova.virt.hardware [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 718.841437] env[63372]: DEBUG nova.virt.hardware [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 718.841565] env[63372]: DEBUG nova.virt.hardware [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.842721] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66cdc47-2f34-4ba8-a8ce-841557239f03 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.855307] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cdc413e-2897-43f6-bc35-0bb5478620eb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.894140] env[63372]: DEBUG nova.scheduler.client.report [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.993586] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d2b3597-0461-41d0-bfeb-39b95646471b tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.976s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.174198] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.184737] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquiring lock "refresh_cache-c825d2e2-a9cc-4c7a-b92e-039756d4121d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.184972] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquired lock "refresh_cache-c825d2e2-a9cc-4c7a-b92e-039756d4121d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.185117] env[63372]: DEBUG nova.network.neutron [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 719.264842] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023855, 'name': RelocateVM_Task} progress is 95%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.264842] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023859, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.400263] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.618s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.400762] env[63372]: DEBUG nova.compute.manager [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 719.403405] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.070s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.404935] env[63372]: INFO nova.compute.claims [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 719.528840] env[63372]: DEBUG nova.network.neutron [req-4d91feaa-bcf7-4c2d-8f29-0b3683a46711 req-86b78ffe-f5b7-48a7-a388-2a5a7ed126d9 service nova] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Updated VIF entry in instance network info cache for port 022ca5fc-1544-4248-bb33-e377ab46b598. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 719.529263] env[63372]: DEBUG nova.network.neutron [req-4d91feaa-bcf7-4c2d-8f29-0b3683a46711 req-86b78ffe-f5b7-48a7-a388-2a5a7ed126d9 service nova] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Updating instance_info_cache with network_info: [{"id": "022ca5fc-1544-4248-bb33-e377ab46b598", "address": "fa:16:3e:57:ea:33", "network": {"id": "b924481a-4e51-4483-a710-f01fc39c851c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1289474667-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b6afb5579f2b48e4b094f6490136417f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap022ca5fc-15", "ovs_interfaceid": "022ca5fc-1544-4248-bb33-e377ab46b598", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.570461] env[63372]: DEBUG nova.network.neutron [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Successfully updated port: c56eef25-45d9-459d-b839-c069b18a613c {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 719.741524] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023855, 'name': RelocateVM_Task} progress is 97%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.751797] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023859, 'name': CreateVM_Task, 'duration_secs': 0.695824} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.752057] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 719.753036] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.753036] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.753385] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 719.753688] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf5d3928-ebb6-43dc-9904-7445c0090293 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.762176] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 719.762176] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]526278ea-fc8c-9f13-0442-dbb1930d804a" [ 719.762176] env[63372]: _type = "Task" [ 719.762176] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.776209] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526278ea-fc8c-9f13-0442-dbb1930d804a, 'name': SearchDatastore_Task, 'duration_secs': 0.012746} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.776589] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.776822] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 719.777074] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.777224] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.778360] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 719.778360] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed1216fa-e26a-4f56-bb63-9779220f8708 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.788060] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 719.788269] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 719.791255] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-147c8cf4-5107-4583-8e7a-2c914a2a6313 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.797906] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 719.797906] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5201be31-c6fc-139a-056c-4230e196370c" [ 719.797906] env[63372]: _type = "Task" [ 719.797906] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.807453] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5201be31-c6fc-139a-056c-4230e196370c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.909473] env[63372]: DEBUG nova.compute.utils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 719.912953] env[63372]: DEBUG nova.compute.manager [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 719.912953] env[63372]: DEBUG nova.network.neutron [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 719.931341] env[63372]: DEBUG nova.network.neutron [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Updating instance_info_cache with network_info: [{"id": "620d14ec-96a7-4912-bb91-44d2c290d431", "address": "fa:16:3e:f4:07:ec", "network": {"id": "e5ec0f30-c191-4392-95c0-ac48df5cc431", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-828375451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b971b7d7a62d4902a930b10d96cfb51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9a1e09ef-7c9c-45d9-9bf4-55b913524948", "external-id": "nsx-vlan-transportzone-466", "segmentation_id": 466, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap620d14ec-96", "ovs_interfaceid": "620d14ec-96a7-4912-bb91-44d2c290d431", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.978053] env[63372]: DEBUG nova.policy [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b96f55f1271346248fe83e2869aa90e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'de7a1fac8ec941b8b00357f62da80927', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 720.025961] env[63372]: DEBUG nova.compute.manager [req-f3de972f-bee3-43dc-a7ea-b74aa049aeaf req-4d2a5092-0258-46f9-84e6-758840f86b4b service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Received event network-changed-7688e644-e47b-4992-b037-343a71e18d2f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.026180] env[63372]: DEBUG nova.compute.manager [req-f3de972f-bee3-43dc-a7ea-b74aa049aeaf req-4d2a5092-0258-46f9-84e6-758840f86b4b service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Refreshing instance network info cache due to event network-changed-7688e644-e47b-4992-b037-343a71e18d2f. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 720.026451] env[63372]: DEBUG oslo_concurrency.lockutils [req-f3de972f-bee3-43dc-a7ea-b74aa049aeaf req-4d2a5092-0258-46f9-84e6-758840f86b4b service nova] Acquiring lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.026601] env[63372]: DEBUG oslo_concurrency.lockutils [req-f3de972f-bee3-43dc-a7ea-b74aa049aeaf req-4d2a5092-0258-46f9-84e6-758840f86b4b service nova] Acquired lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.026758] env[63372]: DEBUG nova.network.neutron [req-f3de972f-bee3-43dc-a7ea-b74aa049aeaf req-4d2a5092-0258-46f9-84e6-758840f86b4b service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Refreshing network info cache for port 7688e644-e47b-4992-b037-343a71e18d2f {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 720.031574] env[63372]: DEBUG oslo_concurrency.lockutils [req-4d91feaa-bcf7-4c2d-8f29-0b3683a46711 req-86b78ffe-f5b7-48a7-a388-2a5a7ed126d9 service nova] Releasing lock "refresh_cache-63b9095c-fb49-4dc1-a6dc-96529aeaab81" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.073122] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "refresh_cache-b182294d-2de8-4189-af7f-3e2d2c604a8b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.073279] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquired lock "refresh_cache-b182294d-2de8-4189-af7f-3e2d2c604a8b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.073727] env[63372]: DEBUG nova.network.neutron [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 720.244716] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023855, 'name': RelocateVM_Task} progress is 98%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.245964] env[63372]: DEBUG nova.network.neutron [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Successfully created port: c4e27cbd-df52-4e92-9621-3b068191a460 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 720.309642] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5201be31-c6fc-139a-056c-4230e196370c, 'name': SearchDatastore_Task, 'duration_secs': 0.009844} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.310525] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-906a6d71-3de5-4b55-9a9c-991778789d1f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.316339] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 720.316339] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529132a2-a7a2-c0c4-597c-f1d043bc4ab1" [ 720.316339] env[63372]: _type = "Task" [ 720.316339] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.325349] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529132a2-a7a2-c0c4-597c-f1d043bc4ab1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.414416] env[63372]: DEBUG nova.compute.manager [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 720.440996] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Releasing lock "refresh_cache-c825d2e2-a9cc-4c7a-b92e-039756d4121d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.446442] env[63372]: DEBUG nova.compute.manager [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 720.447853] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da28b1e5-5f60-4db7-91be-3c3009f0905d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.610509] env[63372]: DEBUG nova.network.neutron [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.700262] env[63372]: DEBUG nova.compute.manager [req-0f106cf5-9835-4d5a-96fa-651ab0f6ff61 req-90ef55c5-8f70-4164-a13d-1b386ef9f3b0 service nova] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Received event network-vif-plugged-c56eef25-45d9-459d-b839-c069b18a613c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.700262] env[63372]: DEBUG oslo_concurrency.lockutils [req-0f106cf5-9835-4d5a-96fa-651ab0f6ff61 req-90ef55c5-8f70-4164-a13d-1b386ef9f3b0 service nova] Acquiring lock "b182294d-2de8-4189-af7f-3e2d2c604a8b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.700262] env[63372]: DEBUG oslo_concurrency.lockutils [req-0f106cf5-9835-4d5a-96fa-651ab0f6ff61 req-90ef55c5-8f70-4164-a13d-1b386ef9f3b0 service nova] Lock "b182294d-2de8-4189-af7f-3e2d2c604a8b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.700262] env[63372]: DEBUG oslo_concurrency.lockutils [req-0f106cf5-9835-4d5a-96fa-651ab0f6ff61 req-90ef55c5-8f70-4164-a13d-1b386ef9f3b0 service nova] Lock "b182294d-2de8-4189-af7f-3e2d2c604a8b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.700262] env[63372]: DEBUG nova.compute.manager [req-0f106cf5-9835-4d5a-96fa-651ab0f6ff61 req-90ef55c5-8f70-4164-a13d-1b386ef9f3b0 service nova] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] No waiting events found dispatching network-vif-plugged-c56eef25-45d9-459d-b839-c069b18a613c {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 720.700262] env[63372]: WARNING nova.compute.manager [req-0f106cf5-9835-4d5a-96fa-651ab0f6ff61 req-90ef55c5-8f70-4164-a13d-1b386ef9f3b0 service nova] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Received unexpected event network-vif-plugged-c56eef25-45d9-459d-b839-c069b18a613c for instance with vm_state building and task_state spawning. [ 720.700262] env[63372]: DEBUG nova.compute.manager [req-0f106cf5-9835-4d5a-96fa-651ab0f6ff61 req-90ef55c5-8f70-4164-a13d-1b386ef9f3b0 service nova] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Received event network-changed-c56eef25-45d9-459d-b839-c069b18a613c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.700262] env[63372]: DEBUG nova.compute.manager [req-0f106cf5-9835-4d5a-96fa-651ab0f6ff61 req-90ef55c5-8f70-4164-a13d-1b386ef9f3b0 service nova] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Refreshing instance network info cache due to event network-changed-c56eef25-45d9-459d-b839-c069b18a613c. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 720.700262] env[63372]: DEBUG oslo_concurrency.lockutils [req-0f106cf5-9835-4d5a-96fa-651ab0f6ff61 req-90ef55c5-8f70-4164-a13d-1b386ef9f3b0 service nova] Acquiring lock "refresh_cache-b182294d-2de8-4189-af7f-3e2d2c604a8b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.749204] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023855, 'name': RelocateVM_Task, 'duration_secs': 3.863611} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.749204] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Volume attach. Driver type: vmdk {{(pid=63372) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 720.749204] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227234', 'volume_id': 'a483fa3d-7a97-48fa-8fe5-723336b546a9', 'name': 'volume-a483fa3d-7a97-48fa-8fe5-723336b546a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5d819bd2-f7f9-480d-83ae-fbd4d2e0d382', 'attached_at': '', 'detached_at': '', 'volume_id': 'a483fa3d-7a97-48fa-8fe5-723336b546a9', 'serial': 'a483fa3d-7a97-48fa-8fe5-723336b546a9'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 720.750082] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f331d51-6dbe-4eb2-afd6-8318ea958b65 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.782434] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f369c0-a2a8-4c3d-88f2-6967d1c5dc0d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.806836] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] volume-a483fa3d-7a97-48fa-8fe5-723336b546a9/volume-a483fa3d-7a97-48fa-8fe5-723336b546a9.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 720.809886] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b56f985c-83dc-4a42-9e7f-1b4af1055792 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.835514] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529132a2-a7a2-c0c4-597c-f1d043bc4ab1, 'name': SearchDatastore_Task, 'duration_secs': 0.017822} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.841827] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.841827] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 63b9095c-fb49-4dc1-a6dc-96529aeaab81/63b9095c-fb49-4dc1-a6dc-96529aeaab81.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 720.841827] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Waiting for the task: (returnval){ [ 720.841827] env[63372]: value = "task-1023860" [ 720.841827] env[63372]: _type = "Task" [ 720.841827] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.841827] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07e1234d-50e5-4466-9b8a-b5a5af6410a5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.850540] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023860, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.855453] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.859032] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.859032] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.859032] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.859032] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.859032] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 720.859032] env[63372]: value = "task-1023861" [ 720.859032] env[63372]: _type = "Task" [ 720.859032] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.861734] env[63372]: INFO nova.compute.manager [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Terminating instance [ 720.869196] env[63372]: DEBUG nova.compute.manager [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 720.869449] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 720.870250] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a8d020-5a91-4289-9c24-f597e62bfcdf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.873972] env[63372]: DEBUG nova.network.neutron [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Updating instance_info_cache with network_info: [{"id": "c56eef25-45d9-459d-b839-c069b18a613c", "address": "fa:16:3e:7f:1f:ae", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.59", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc56eef25-45", "ovs_interfaceid": "c56eef25-45d9-459d-b839-c069b18a613c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.883059] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023861, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.885579] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 720.885832] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8df86cd3-c3c0-43ce-afab-230561c88136 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.893523] env[63372]: DEBUG oslo_vmware.api [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Waiting for the task: (returnval){ [ 720.893523] env[63372]: value = "task-1023862" [ 720.893523] env[63372]: _type = "Task" [ 720.893523] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.904617] env[63372]: DEBUG oslo_vmware.api [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023862, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.917404] env[63372]: DEBUG nova.network.neutron [req-f3de972f-bee3-43dc-a7ea-b74aa049aeaf req-4d2a5092-0258-46f9-84e6-758840f86b4b service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Updated VIF entry in instance network info cache for port 7688e644-e47b-4992-b037-343a71e18d2f. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 720.917404] env[63372]: DEBUG nova.network.neutron [req-f3de972f-bee3-43dc-a7ea-b74aa049aeaf req-4d2a5092-0258-46f9-84e6-758840f86b4b service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Updating instance_info_cache with network_info: [{"id": "7688e644-e47b-4992-b037-343a71e18d2f", "address": "fa:16:3e:dc:d7:8d", "network": {"id": "d26c71aa-f3c4-4b1a-a263-fa95047e2b74", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1927982377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b68c21179fb24008baf1da0d0119fb29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6e2a9679-6746-40f2-951c-65fcd1af5f7b", "external-id": "nsx-vlan-transportzone-39", "segmentation_id": 39, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7688e644-e4", "ovs_interfaceid": "7688e644-e47b-4992-b037-343a71e18d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.944142] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908ec36c-86eb-400d-98ad-24aecae2d7fc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.952336] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf276308-6e5c-4778-9a6a-3d9a6f88e460 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.994863] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c098f34f-9b34-4a29-8d40-78c694fea5e6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.004342] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f523c5f3-bc83-4bfe-9f7c-42bf7abc441e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.020884] env[63372]: DEBUG nova.compute.provider_tree [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.353841] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023860, 'name': ReconfigVM_Task, 'duration_secs': 0.412665} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.353841] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Reconfigured VM instance instance-0000002b to attach disk [datastore1] volume-a483fa3d-7a97-48fa-8fe5-723336b546a9/volume-a483fa3d-7a97-48fa-8fe5-723336b546a9.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 721.358379] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-003a07ea-b222-4d1e-b618-3b4684084c38 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.380137] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Releasing lock "refresh_cache-b182294d-2de8-4189-af7f-3e2d2c604a8b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.380511] env[63372]: DEBUG nova.compute.manager [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Instance network_info: |[{"id": "c56eef25-45d9-459d-b839-c069b18a613c", "address": "fa:16:3e:7f:1f:ae", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.59", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc56eef25-45", "ovs_interfaceid": "c56eef25-45d9-459d-b839-c069b18a613c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 721.380809] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023861, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.382283] env[63372]: DEBUG oslo_concurrency.lockutils [req-0f106cf5-9835-4d5a-96fa-651ab0f6ff61 req-90ef55c5-8f70-4164-a13d-1b386ef9f3b0 service nova] Acquired lock "refresh_cache-b182294d-2de8-4189-af7f-3e2d2c604a8b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.382481] env[63372]: DEBUG nova.network.neutron [req-0f106cf5-9835-4d5a-96fa-651ab0f6ff61 req-90ef55c5-8f70-4164-a13d-1b386ef9f3b0 service nova] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Refreshing network info cache for port c56eef25-45d9-459d-b839-c069b18a613c {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 721.383759] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:1f:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c56eef25-45d9-459d-b839-c069b18a613c', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 721.392146] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Creating folder: Project (de7a1fac8ec941b8b00357f62da80927). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 721.392861] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Waiting for the task: (returnval){ [ 721.392861] env[63372]: value = "task-1023863" [ 721.392861] env[63372]: _type = "Task" [ 721.392861] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.393171] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-95260430-84e6-4585-8a37-e116d9661050 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.415046] env[63372]: DEBUG oslo_vmware.api [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023862, 'name': PowerOffVM_Task, 'duration_secs': 0.20946} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.418771] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 721.418997] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 721.420390] env[63372]: DEBUG oslo_concurrency.lockutils [req-f3de972f-bee3-43dc-a7ea-b74aa049aeaf req-4d2a5092-0258-46f9-84e6-758840f86b4b service nova] Releasing lock "refresh_cache-e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.420390] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023863, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.421867] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c4e63f11-7ea9-4a1d-a0ec-20b61cda4048 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.423722] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Created folder: Project (de7a1fac8ec941b8b00357f62da80927) in parent group-v227230. [ 721.423928] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Creating folder: Instances. Parent ref: group-v227300. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 721.424194] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eda1d163-8155-4e5c-8ac0-e34b3135fee3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.428584] env[63372]: DEBUG nova.compute.manager [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 721.440705] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Created folder: Instances in parent group-v227300. [ 721.440705] env[63372]: DEBUG oslo.service.loopingcall [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 721.441200] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 721.442746] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-19e11d2a-ac91-49c2-aabe-33236c35b473 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.465734] env[63372]: DEBUG nova.virt.hardware [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 721.466178] env[63372]: DEBUG nova.virt.hardware [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 721.466178] env[63372]: DEBUG nova.virt.hardware [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 721.466376] env[63372]: DEBUG nova.virt.hardware [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 721.466756] env[63372]: DEBUG nova.virt.hardware [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 721.466756] env[63372]: DEBUG nova.virt.hardware [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 721.467018] env[63372]: DEBUG nova.virt.hardware [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 721.467018] env[63372]: DEBUG nova.virt.hardware [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 721.467314] env[63372]: DEBUG nova.virt.hardware [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 721.467381] env[63372]: DEBUG nova.virt.hardware [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 721.467611] env[63372]: DEBUG nova.virt.hardware [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 721.468560] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5619861-7f0f-49f0-9286-c592c2cc19a5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.473446] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 721.473446] env[63372]: value = "task-1023867" [ 721.473446] env[63372]: _type = "Task" [ 721.473446] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.480706] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8e383e9-144c-4e0e-8814-5b72b24fd41c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.488329] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023867, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.501262] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23152ffb-2c9f-478f-a23d-35e5e3dd7a60 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.511013] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Doing hard reboot of VM {{(pid=63372) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 721.511013] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-36fda770-6799-4892-bdd7-9a565c7cf1a1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.518472] env[63372]: DEBUG oslo_vmware.api [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Waiting for the task: (returnval){ [ 721.518472] env[63372]: value = "task-1023868" [ 721.518472] env[63372]: _type = "Task" [ 721.518472] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.524424] env[63372]: DEBUG nova.scheduler.client.report [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.534382] env[63372]: DEBUG oslo_vmware.api [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023868, 'name': ResetVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.879390] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023861, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.614893} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.879922] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 63b9095c-fb49-4dc1-a6dc-96529aeaab81/63b9095c-fb49-4dc1-a6dc-96529aeaab81.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 721.879922] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 721.880104] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8837ad36-4de3-45d9-b95e-4caa9a075391 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.888051] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 721.888051] env[63372]: value = "task-1023869" [ 721.888051] env[63372]: _type = "Task" [ 721.888051] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.900187] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023869, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.909980] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023863, 'name': ReconfigVM_Task, 'duration_secs': 0.181401} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.912022] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227234', 'volume_id': 'a483fa3d-7a97-48fa-8fe5-723336b546a9', 'name': 'volume-a483fa3d-7a97-48fa-8fe5-723336b546a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5d819bd2-f7f9-480d-83ae-fbd4d2e0d382', 'attached_at': '', 'detached_at': '', 'volume_id': 'a483fa3d-7a97-48fa-8fe5-723336b546a9', 'serial': 'a483fa3d-7a97-48fa-8fe5-723336b546a9'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 721.912022] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f42c2b27-cff1-4016-81b5-b642ceebca51 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.917748] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Waiting for the task: (returnval){ [ 721.917748] env[63372]: value = "task-1023870" [ 721.917748] env[63372]: _type = "Task" [ 721.917748] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.929489] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023870, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.986052] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023867, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.030957] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.627s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.030957] env[63372]: DEBUG nova.compute.manager [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 722.033431] env[63372]: DEBUG oslo_vmware.api [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023868, 'name': ResetVM_Task, 'duration_secs': 0.110866} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.033939] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.116s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.039580] env[63372]: INFO nova.compute.claims [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 722.042170] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Did hard reboot of VM {{(pid=63372) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 722.042350] env[63372]: DEBUG nova.compute.manager [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 722.046572] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7e35ef-9773-48e2-8f63-15e5410ec3bf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.110574] env[63372]: DEBUG nova.network.neutron [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Successfully updated port: c4e27cbd-df52-4e92-9621-3b068191a460 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 722.197833] env[63372]: DEBUG nova.network.neutron [req-0f106cf5-9835-4d5a-96fa-651ab0f6ff61 req-90ef55c5-8f70-4164-a13d-1b386ef9f3b0 service nova] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Updated VIF entry in instance network info cache for port c56eef25-45d9-459d-b839-c069b18a613c. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 722.198215] env[63372]: DEBUG nova.network.neutron [req-0f106cf5-9835-4d5a-96fa-651ab0f6ff61 req-90ef55c5-8f70-4164-a13d-1b386ef9f3b0 service nova] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Updating instance_info_cache with network_info: [{"id": "c56eef25-45d9-459d-b839-c069b18a613c", "address": "fa:16:3e:7f:1f:ae", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.59", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc56eef25-45", "ovs_interfaceid": "c56eef25-45d9-459d-b839-c069b18a613c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.397935] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023869, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.213319} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.398235] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 722.399029] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d38c2461-6e2d-4ce4-96e8-64c39880526c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.421416] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] 63b9095c-fb49-4dc1-a6dc-96529aeaab81/63b9095c-fb49-4dc1-a6dc-96529aeaab81.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 722.421712] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6edd7f32-b8fb-40a0-8215-798bb2312f0f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.445425] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023870, 'name': Rename_Task, 'duration_secs': 0.29609} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.446652] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 722.446951] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 722.446951] env[63372]: value = "task-1023871" [ 722.446951] env[63372]: _type = "Task" [ 722.446951] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.447141] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b207d1fe-4d57-4130-842b-b675c4382b60 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.456926] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023871, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.458141] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Waiting for the task: (returnval){ [ 722.458141] env[63372]: value = "task-1023872" [ 722.458141] env[63372]: _type = "Task" [ 722.458141] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.465651] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023872, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.483675] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023867, 'name': CreateVM_Task, 'duration_secs': 0.765728} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.483850] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 722.484574] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.484746] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.485066] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 722.485305] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e1ec636-25bc-4004-a9b5-2ef063741c37 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.489746] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 722.489746] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f1824b-80d0-0167-598f-9f618ab3c759" [ 722.489746] env[63372]: _type = "Task" [ 722.489746] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.497785] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f1824b-80d0-0167-598f-9f618ab3c759, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.535390] env[63372]: DEBUG nova.compute.utils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 722.538180] env[63372]: DEBUG nova.compute.manager [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 722.538453] env[63372]: DEBUG nova.network.neutron [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 722.540238] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 722.540547] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 722.540737] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Deleting the datastore file [datastore1] e5522005-8c54-43e1-ae23-5e5ff1ef0ee9 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 722.541224] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b33a440-cea3-49f6-9af1-f486002a6fc2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.549900] env[63372]: DEBUG oslo_vmware.api [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Waiting for the task: (returnval){ [ 722.549900] env[63372]: value = "task-1023873" [ 722.549900] env[63372]: _type = "Task" [ 722.549900] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.560459] env[63372]: DEBUG oslo_vmware.api [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023873, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.565026] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf251a34-f63a-4ae7-a64f-b328a5d5e208 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.910s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.613021] env[63372]: DEBUG nova.policy [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1048d032dc00441f9c2f99b60f357dfb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5cf13f5cc32c4b3ebaee47f03bb64ba7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 722.613432] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "refresh_cache-f0c60559-c072-4b61-afe8-03d6c131b307" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.613579] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquired lock "refresh_cache-f0c60559-c072-4b61-afe8-03d6c131b307" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.613724] env[63372]: DEBUG nova.network.neutron [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 722.701792] env[63372]: DEBUG oslo_concurrency.lockutils [req-0f106cf5-9835-4d5a-96fa-651ab0f6ff61 req-90ef55c5-8f70-4164-a13d-1b386ef9f3b0 service nova] Releasing lock "refresh_cache-b182294d-2de8-4189-af7f-3e2d2c604a8b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.723814] env[63372]: DEBUG nova.compute.manager [req-00051577-92f8-4749-8769-b9d57d4424d1 req-42bf2aa1-b8cd-4ae2-bfac-4c088cbd76f4 service nova] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Received event network-vif-plugged-c4e27cbd-df52-4e92-9621-3b068191a460 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.724042] env[63372]: DEBUG oslo_concurrency.lockutils [req-00051577-92f8-4749-8769-b9d57d4424d1 req-42bf2aa1-b8cd-4ae2-bfac-4c088cbd76f4 service nova] Acquiring lock "f0c60559-c072-4b61-afe8-03d6c131b307-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.724249] env[63372]: DEBUG oslo_concurrency.lockutils [req-00051577-92f8-4749-8769-b9d57d4424d1 req-42bf2aa1-b8cd-4ae2-bfac-4c088cbd76f4 service nova] Lock "f0c60559-c072-4b61-afe8-03d6c131b307-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.724458] env[63372]: DEBUG oslo_concurrency.lockutils [req-00051577-92f8-4749-8769-b9d57d4424d1 req-42bf2aa1-b8cd-4ae2-bfac-4c088cbd76f4 service nova] Lock "f0c60559-c072-4b61-afe8-03d6c131b307-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.724570] env[63372]: DEBUG nova.compute.manager [req-00051577-92f8-4749-8769-b9d57d4424d1 req-42bf2aa1-b8cd-4ae2-bfac-4c088cbd76f4 service nova] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] No waiting events found dispatching network-vif-plugged-c4e27cbd-df52-4e92-9621-3b068191a460 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 722.724731] env[63372]: WARNING nova.compute.manager [req-00051577-92f8-4749-8769-b9d57d4424d1 req-42bf2aa1-b8cd-4ae2-bfac-4c088cbd76f4 service nova] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Received unexpected event network-vif-plugged-c4e27cbd-df52-4e92-9621-3b068191a460 for instance with vm_state building and task_state spawning. [ 722.724886] env[63372]: DEBUG nova.compute.manager [req-00051577-92f8-4749-8769-b9d57d4424d1 req-42bf2aa1-b8cd-4ae2-bfac-4c088cbd76f4 service nova] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Received event network-changed-c4e27cbd-df52-4e92-9621-3b068191a460 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.725297] env[63372]: DEBUG nova.compute.manager [req-00051577-92f8-4749-8769-b9d57d4424d1 req-42bf2aa1-b8cd-4ae2-bfac-4c088cbd76f4 service nova] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Refreshing instance network info cache due to event network-changed-c4e27cbd-df52-4e92-9621-3b068191a460. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 722.725477] env[63372]: DEBUG oslo_concurrency.lockutils [req-00051577-92f8-4749-8769-b9d57d4424d1 req-42bf2aa1-b8cd-4ae2-bfac-4c088cbd76f4 service nova] Acquiring lock "refresh_cache-f0c60559-c072-4b61-afe8-03d6c131b307" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.864116] env[63372]: DEBUG nova.network.neutron [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Successfully created port: 874f14b1-6cc3-4701-8a08-35940fc9bd9b {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.958921] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.965829] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023872, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.999369] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f1824b-80d0-0167-598f-9f618ab3c759, 'name': SearchDatastore_Task, 'duration_secs': 0.010596} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.999655] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.999882] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 723.000117] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.000266] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.000436] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 723.000684] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44f1565e-4a44-43e7-8892-a4f887854f8f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.008603] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 723.008783] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 723.009482] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4727876e-fb6c-4037-adff-ce3e0a6e7e7f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.014371] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 723.014371] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523e57de-6e0b-8da7-404d-10421e426b98" [ 723.014371] env[63372]: _type = "Task" [ 723.014371] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.022596] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523e57de-6e0b-8da7-404d-10421e426b98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.042442] env[63372]: DEBUG nova.compute.manager [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 723.058734] env[63372]: DEBUG oslo_vmware.api [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Task: {'id': task-1023873, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152493} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.058872] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 723.059060] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 723.059231] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 723.059403] env[63372]: INFO nova.compute.manager [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Took 2.19 seconds to destroy the instance on the hypervisor. [ 723.059637] env[63372]: DEBUG oslo.service.loopingcall [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 723.059822] env[63372]: DEBUG nova.compute.manager [-] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 723.059914] env[63372]: DEBUG nova.network.neutron [-] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 723.173817] env[63372]: DEBUG nova.network.neutron [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.416186] env[63372]: DEBUG nova.network.neutron [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Updating instance_info_cache with network_info: [{"id": "c4e27cbd-df52-4e92-9621-3b068191a460", "address": "fa:16:3e:c4:d0:17", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.23", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e27cbd-df", "ovs_interfaceid": "c4e27cbd-df52-4e92-9621-3b068191a460", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.461193] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023871, 'name': ReconfigVM_Task, 'duration_secs': 0.586185} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.466546] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Reconfigured VM instance instance-0000002c to attach disk [datastore1] 63b9095c-fb49-4dc1-a6dc-96529aeaab81/63b9095c-fb49-4dc1-a6dc-96529aeaab81.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 723.468397] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ba47aa45-8c6e-4101-a478-f45f9faa7353 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.473815] env[63372]: DEBUG oslo_vmware.api [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1023872, 'name': PowerOnVM_Task, 'duration_secs': 0.697514} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.475199] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 723.475286] env[63372]: INFO nova.compute.manager [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Took 7.95 seconds to spawn the instance on the hypervisor. [ 723.476327] env[63372]: DEBUG nova.compute.manager [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 723.476327] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 723.476327] env[63372]: value = "task-1023874" [ 723.476327] env[63372]: _type = "Task" [ 723.476327] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.476811] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9177618-a983-4b5e-ac95-ea0d2c2c02ed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.496831] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023874, 'name': Rename_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.516418] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e435986-661b-4cee-a1a3-fdee5222a2ca {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.532020] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523e57de-6e0b-8da7-404d-10421e426b98, 'name': SearchDatastore_Task, 'duration_secs': 0.008895} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.532020] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2d1850-607f-44fb-b0cc-f0ea220c7f82 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.535063] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8d906e5-fd1a-40a6-96dc-5e04bfc893ff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.569586] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3166296-9999-448d-9907-7a80f2e9ce94 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.572603] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 723.572603] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]522e388d-eb0d-779c-4946-77389b7553ba" [ 723.572603] env[63372]: _type = "Task" [ 723.572603] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.582096] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ed141c-f9c8-4f5e-88e9-2e2428ecc85b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.588023] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522e388d-eb0d-779c-4946-77389b7553ba, 'name': SearchDatastore_Task, 'duration_secs': 0.010942} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.588661] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.589024] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] b182294d-2de8-4189-af7f-3e2d2c604a8b/b182294d-2de8-4189-af7f-3e2d2c604a8b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 723.589313] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-991c83a2-88f1-4929-894b-7f1d3c2c8ece {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.599269] env[63372]: DEBUG nova.compute.provider_tree [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.605591] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 723.605591] env[63372]: value = "task-1023875" [ 723.605591] env[63372]: _type = "Task" [ 723.605591] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.615779] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023875, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.880549] env[63372]: DEBUG nova.network.neutron [-] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.921976] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Releasing lock "refresh_cache-f0c60559-c072-4b61-afe8-03d6c131b307" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.921976] env[63372]: DEBUG nova.compute.manager [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Instance network_info: |[{"id": "c4e27cbd-df52-4e92-9621-3b068191a460", "address": "fa:16:3e:c4:d0:17", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.23", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e27cbd-df", "ovs_interfaceid": "c4e27cbd-df52-4e92-9621-3b068191a460", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 723.921976] env[63372]: DEBUG oslo_concurrency.lockutils [req-00051577-92f8-4749-8769-b9d57d4424d1 req-42bf2aa1-b8cd-4ae2-bfac-4c088cbd76f4 service nova] Acquired lock "refresh_cache-f0c60559-c072-4b61-afe8-03d6c131b307" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.921976] env[63372]: DEBUG nova.network.neutron [req-00051577-92f8-4749-8769-b9d57d4424d1 req-42bf2aa1-b8cd-4ae2-bfac-4c088cbd76f4 service nova] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Refreshing network info cache for port c4e27cbd-df52-4e92-9621-3b068191a460 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 723.921976] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:d0:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c4e27cbd-df52-4e92-9621-3b068191a460', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 723.936920] env[63372]: DEBUG oslo.service.loopingcall [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 723.936920] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 723.936920] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04e11111-d7b3-4d6e-a3b2-0887d49f0960 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.957272] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 723.957272] env[63372]: value = "task-1023876" [ 723.957272] env[63372]: _type = "Task" [ 723.957272] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.965979] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023876, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.989312] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023874, 'name': Rename_Task, 'duration_secs': 0.208655} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.989612] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 723.990113] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-83b156cf-aff2-4fdb-a311-ff454cf9e4b1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.999097] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 723.999097] env[63372]: value = "task-1023877" [ 723.999097] env[63372]: _type = "Task" [ 723.999097] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.005454] env[63372]: INFO nova.compute.manager [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Took 45.11 seconds to build instance. [ 724.012764] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023877, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.071274] env[63372]: DEBUG nova.compute.manager [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 724.103685] env[63372]: DEBUG nova.virt.hardware [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:36:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='51633402-95cc-485c-8808-65539d485326',id=27,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-443337402',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 724.103972] env[63372]: DEBUG nova.virt.hardware [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 724.104209] env[63372]: DEBUG nova.virt.hardware [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 724.104565] env[63372]: DEBUG nova.virt.hardware [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 724.104720] env[63372]: DEBUG nova.virt.hardware [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 724.104866] env[63372]: DEBUG nova.virt.hardware [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 724.105086] env[63372]: DEBUG nova.virt.hardware [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 724.105289] env[63372]: DEBUG nova.virt.hardware [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 724.109019] env[63372]: DEBUG nova.virt.hardware [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 724.109019] env[63372]: DEBUG nova.virt.hardware [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 724.109019] env[63372]: DEBUG nova.virt.hardware [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 724.109019] env[63372]: DEBUG nova.scheduler.client.report [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 724.110710] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b7af05-8938-4da3-8db3-0b1e0f2edc84 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.126307] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9813608d-cfd8-4fd4-9116-451255d11d52 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.131082] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023875, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517101} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.132225] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] b182294d-2de8-4189-af7f-3e2d2c604a8b/b182294d-2de8-4189-af7f-3e2d2c604a8b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 724.132493] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 724.133177] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-768371a4-04f7-463f-bf9a-07b6287fe81d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.149494] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 724.149494] env[63372]: value = "task-1023878" [ 724.149494] env[63372]: _type = "Task" [ 724.149494] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.157749] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023878, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.192327] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquiring lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.192327] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.192327] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquiring lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.192327] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.192327] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.196500] env[63372]: INFO nova.compute.manager [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Terminating instance [ 724.199136] env[63372]: DEBUG nova.compute.manager [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 724.203023] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 724.203023] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3ca93f-b777-4a16-90c4-543add32aa40 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.209190] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 724.209468] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eec5ba78-ea00-423d-8178-ec7c4a72c3a9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.218553] env[63372]: DEBUG oslo_vmware.api [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Waiting for the task: (returnval){ [ 724.218553] env[63372]: value = "task-1023879" [ 724.218553] env[63372]: _type = "Task" [ 724.218553] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.230472] env[63372]: DEBUG oslo_vmware.api [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023879, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.384225] env[63372]: INFO nova.compute.manager [-] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Took 1.32 seconds to deallocate network for instance. [ 724.467859] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023876, 'name': CreateVM_Task, 'duration_secs': 0.434375} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.468627] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 724.469637] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.469926] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.470353] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 724.472215] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-645e38ca-3b53-4386-922b-d5bf4d57a12f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.479342] env[63372]: DEBUG nova.network.neutron [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Successfully updated port: 874f14b1-6cc3-4701-8a08-35940fc9bd9b {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 724.487324] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 724.487324] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]527175fa-d41d-c5a6-4890-4c1763ec632d" [ 724.487324] env[63372]: _type = "Task" [ 724.487324] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.496431] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527175fa-d41d-c5a6-4890-4c1763ec632d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.512092] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0f48a618-860d-4a2e-b969-a48ee0f6380f tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Lock "5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.958s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.513101] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023877, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.618020] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.581s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.618020] env[63372]: DEBUG nova.compute.manager [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 724.619723] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.401s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.620589] env[63372]: DEBUG nova.objects.instance [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Lazy-loading 'resources' on Instance uuid 276c3490-240a-499b-ba64-4b02510b169c {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 724.668394] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023878, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097673} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.668839] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 724.670290] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372fa8bc-e33d-4d2f-a421-298ffa03b23f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.698263] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Reconfiguring VM instance instance-0000002d to attach disk [datastore1] b182294d-2de8-4189-af7f-3e2d2c604a8b/b182294d-2de8-4189-af7f-3e2d2c604a8b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 724.698582] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7732c8f8-ea63-4510-b715-910b92751584 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.724691] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 724.724691] env[63372]: value = "task-1023880" [ 724.724691] env[63372]: _type = "Task" [ 724.724691] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.730443] env[63372]: DEBUG oslo_vmware.api [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023879, 'name': PowerOffVM_Task, 'duration_secs': 0.219538} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.731077] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 724.731456] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 724.731536] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf1eab9e-3ed3-44f0-a3e5-ddc98ec020eb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.735804] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023880, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.802708] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 724.802936] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 724.803135] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Deleting the datastore file [datastore1] c825d2e2-a9cc-4c7a-b92e-039756d4121d {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 724.803407] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-119ece37-1560-4191-ac74-316973bf58b6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.812503] env[63372]: DEBUG oslo_vmware.api [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Waiting for the task: (returnval){ [ 724.812503] env[63372]: value = "task-1023882" [ 724.812503] env[63372]: _type = "Task" [ 724.812503] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.818559] env[63372]: DEBUG oslo_vmware.api [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023882, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.896312] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.986272] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.986409] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquired lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.986555] env[63372]: DEBUG nova.network.neutron [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 724.998697] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527175fa-d41d-c5a6-4890-4c1763ec632d, 'name': SearchDatastore_Task, 'duration_secs': 0.009282} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.000170] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.000413] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 725.000653] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.000797] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.000983] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 725.007289] env[63372]: DEBUG nova.compute.manager [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Received event network-vif-deleted-7688e644-e47b-4992-b037-343a71e18d2f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.007289] env[63372]: DEBUG nova.compute.manager [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Received event network-vif-plugged-874f14b1-6cc3-4701-8a08-35940fc9bd9b {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.007289] env[63372]: DEBUG oslo_concurrency.lockutils [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] Acquiring lock "8283b736-ad02-4082-97b7-561bd5c5da93-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.007289] env[63372]: DEBUG oslo_concurrency.lockutils [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] Lock "8283b736-ad02-4082-97b7-561bd5c5da93-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.007289] env[63372]: DEBUG oslo_concurrency.lockutils [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] Lock "8283b736-ad02-4082-97b7-561bd5c5da93-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.007289] env[63372]: DEBUG nova.compute.manager [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] No waiting events found dispatching network-vif-plugged-874f14b1-6cc3-4701-8a08-35940fc9bd9b {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 725.007289] env[63372]: WARNING nova.compute.manager [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Received unexpected event network-vif-plugged-874f14b1-6cc3-4701-8a08-35940fc9bd9b for instance with vm_state building and task_state spawning. [ 725.007289] env[63372]: DEBUG nova.compute.manager [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Received event network-changed-874f14b1-6cc3-4701-8a08-35940fc9bd9b {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.007289] env[63372]: DEBUG nova.compute.manager [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Refreshing instance network info cache due to event network-changed-874f14b1-6cc3-4701-8a08-35940fc9bd9b. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 725.007289] env[63372]: DEBUG oslo_concurrency.lockutils [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] Acquiring lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.007289] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d23fbb99-3a71-490d-a084-638c77e6a67e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.015990] env[63372]: DEBUG nova.compute.manager [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 725.023220] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 725.023429] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 725.024287] env[63372]: DEBUG oslo_vmware.api [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023877, 'name': PowerOnVM_Task, 'duration_secs': 0.993631} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.024956] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8836b0d0-6abc-4656-bc26-fbcf75ef9c8e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.028401] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 725.028524] env[63372]: INFO nova.compute.manager [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Took 8.79 seconds to spawn the instance on the hypervisor. [ 725.028698] env[63372]: DEBUG nova.compute.manager [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 725.029501] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d63eb6d2-d2b4-47a6-8595-065aed4fd9aa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.036213] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 725.036213] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52790972-88e9-4fb7-e34b-bc4f43cd1bf1" [ 725.036213] env[63372]: _type = "Task" [ 725.036213] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.051364] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52790972-88e9-4fb7-e34b-bc4f43cd1bf1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.103022] env[63372]: DEBUG nova.network.neutron [req-00051577-92f8-4749-8769-b9d57d4424d1 req-42bf2aa1-b8cd-4ae2-bfac-4c088cbd76f4 service nova] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Updated VIF entry in instance network info cache for port c4e27cbd-df52-4e92-9621-3b068191a460. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 725.103397] env[63372]: DEBUG nova.network.neutron [req-00051577-92f8-4749-8769-b9d57d4424d1 req-42bf2aa1-b8cd-4ae2-bfac-4c088cbd76f4 service nova] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Updating instance_info_cache with network_info: [{"id": "c4e27cbd-df52-4e92-9621-3b068191a460", "address": "fa:16:3e:c4:d0:17", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.23", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e27cbd-df", "ovs_interfaceid": "c4e27cbd-df52-4e92-9621-3b068191a460", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.123742] env[63372]: DEBUG nova.compute.utils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 725.127599] env[63372]: DEBUG nova.compute.manager [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 725.127836] env[63372]: DEBUG nova.network.neutron [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 725.216907] env[63372]: DEBUG nova.policy [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d822682f06640489a48f8577acf4389', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e46b6acae4a4b71bc390bd513ddd2f1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 725.239896] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023880, 'name': ReconfigVM_Task, 'duration_secs': 0.425526} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.240635] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Reconfigured VM instance instance-0000002d to attach disk [datastore1] b182294d-2de8-4189-af7f-3e2d2c604a8b/b182294d-2de8-4189-af7f-3e2d2c604a8b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 725.241611] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0ad9080-fbfb-4a30-9432-30454ca03174 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.253027] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 725.253027] env[63372]: value = "task-1023883" [ 725.253027] env[63372]: _type = "Task" [ 725.253027] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.262275] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023883, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.319833] env[63372]: DEBUG oslo_vmware.api [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Task: {'id': task-1023882, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127083} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.323259] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 725.323259] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 725.323259] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 725.323259] env[63372]: INFO nova.compute.manager [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 725.323876] env[63372]: DEBUG oslo.service.loopingcall [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.328310] env[63372]: DEBUG nova.compute.manager [-] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.328310] env[63372]: DEBUG nova.network.neutron [-] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 725.503484] env[63372]: DEBUG nova.compute.manager [req-1927776b-c8c3-4ca7-aab4-ab7e04031154 req-1caf1040-ed86-4c2f-ac71-6296360008a9 service nova] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Received event network-changed-62ba51bf-2e6c-4205-93c0-e45bfd98f6ae {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.503766] env[63372]: DEBUG nova.compute.manager [req-1927776b-c8c3-4ca7-aab4-ab7e04031154 req-1caf1040-ed86-4c2f-ac71-6296360008a9 service nova] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Refreshing instance network info cache due to event network-changed-62ba51bf-2e6c-4205-93c0-e45bfd98f6ae. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 725.503884] env[63372]: DEBUG oslo_concurrency.lockutils [req-1927776b-c8c3-4ca7-aab4-ab7e04031154 req-1caf1040-ed86-4c2f-ac71-6296360008a9 service nova] Acquiring lock "refresh_cache-5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.504035] env[63372]: DEBUG oslo_concurrency.lockutils [req-1927776b-c8c3-4ca7-aab4-ab7e04031154 req-1caf1040-ed86-4c2f-ac71-6296360008a9 service nova] Acquired lock "refresh_cache-5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.504591] env[63372]: DEBUG nova.network.neutron [req-1927776b-c8c3-4ca7-aab4-ab7e04031154 req-1caf1040-ed86-4c2f-ac71-6296360008a9 service nova] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Refreshing network info cache for port 62ba51bf-2e6c-4205-93c0-e45bfd98f6ae {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 725.542325] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.546970] env[63372]: DEBUG nova.network.neutron [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.561084] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52790972-88e9-4fb7-e34b-bc4f43cd1bf1, 'name': SearchDatastore_Task, 'duration_secs': 0.039041} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.561655] env[63372]: INFO nova.compute.manager [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Took 37.19 seconds to build instance. [ 725.563408] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-451d44b6-fab2-46f6-a58b-8e76faf1fe91 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.569019] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 725.569019] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5211b36a-e568-a93f-2110-ff71c76715bb" [ 725.569019] env[63372]: _type = "Task" [ 725.569019] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.584771] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5211b36a-e568-a93f-2110-ff71c76715bb, 'name': SearchDatastore_Task, 'duration_secs': 0.009894} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.585027] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.585214] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] f0c60559-c072-4b61-afe8-03d6c131b307/f0c60559-c072-4b61-afe8-03d6c131b307.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 725.586274] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7298a44b-d805-496c-af8d-d85b7be4885a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.594205] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 725.594205] env[63372]: value = "task-1023884" [ 725.594205] env[63372]: _type = "Task" [ 725.594205] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.604304] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023884, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.607200] env[63372]: DEBUG oslo_concurrency.lockutils [req-00051577-92f8-4749-8769-b9d57d4424d1 req-42bf2aa1-b8cd-4ae2-bfac-4c088cbd76f4 service nova] Releasing lock "refresh_cache-f0c60559-c072-4b61-afe8-03d6c131b307" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.631327] env[63372]: DEBUG nova.compute.manager [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 725.641387] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d077c2-9887-47a9-afd3-9eda2955470c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.654606] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349054d1-12d7-4e0c-b711-a086a0afadf3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.688728] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7aec915-fb75-4a9c-b901-22d776fabbfa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.698128] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-011af42c-0d09-4940-afed-887c49b39bc1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.714099] env[63372]: DEBUG nova.compute.provider_tree [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.763639] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023883, 'name': Rename_Task, 'duration_secs': 0.267704} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.763796] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 725.764015] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5f34364-841a-4757-b39a-786942191873 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.770909] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 725.770909] env[63372]: value = "task-1023885" [ 725.770909] env[63372]: _type = "Task" [ 725.770909] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.779726] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023885, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.780795] env[63372]: DEBUG nova.network.neutron [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance_info_cache with network_info: [{"id": "874f14b1-6cc3-4701-8a08-35940fc9bd9b", "address": "fa:16:3e:57:46:8e", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.24", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap874f14b1-6c", "ovs_interfaceid": "874f14b1-6cc3-4701-8a08-35940fc9bd9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.806191] env[63372]: DEBUG nova.network.neutron [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Successfully created port: 8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 726.065392] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3cf49430-934e-4f7b-bfe2-9903112c350a tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "63b9095c-fb49-4dc1-a6dc-96529aeaab81" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.908s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.112270] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023884, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475511} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.112270] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] f0c60559-c072-4b61-afe8-03d6c131b307/f0c60559-c072-4b61-afe8-03d6c131b307.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 726.112270] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 726.112270] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e5ff7eed-1ac3-4d2f-937a-34f80996eb44 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.118534] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 726.118534] env[63372]: value = "task-1023886" [ 726.118534] env[63372]: _type = "Task" [ 726.118534] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.126979] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023886, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.193575] env[63372]: DEBUG nova.network.neutron [-] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.219532] env[63372]: DEBUG nova.scheduler.client.report [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.259579] env[63372]: DEBUG nova.network.neutron [req-1927776b-c8c3-4ca7-aab4-ab7e04031154 req-1caf1040-ed86-4c2f-ac71-6296360008a9 service nova] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Updated VIF entry in instance network info cache for port 62ba51bf-2e6c-4205-93c0-e45bfd98f6ae. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 726.260094] env[63372]: DEBUG nova.network.neutron [req-1927776b-c8c3-4ca7-aab4-ab7e04031154 req-1caf1040-ed86-4c2f-ac71-6296360008a9 service nova] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Updating instance_info_cache with network_info: [{"id": "62ba51bf-2e6c-4205-93c0-e45bfd98f6ae", "address": "fa:16:3e:ea:91:de", "network": {"id": "3f4f7142-9adb-40bd-a4c4-50da438a2a82", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-657036764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6dd3c2f3447d41fe8434faa68ad5ea0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62ba51bf-2e", "ovs_interfaceid": "62ba51bf-2e6c-4205-93c0-e45bfd98f6ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.283019] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023885, 'name': PowerOnVM_Task} progress is 79%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.283608] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Releasing lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.284054] env[63372]: DEBUG nova.compute.manager [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Instance network_info: |[{"id": "874f14b1-6cc3-4701-8a08-35940fc9bd9b", "address": "fa:16:3e:57:46:8e", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.24", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap874f14b1-6c", "ovs_interfaceid": "874f14b1-6cc3-4701-8a08-35940fc9bd9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 726.284509] env[63372]: DEBUG oslo_concurrency.lockutils [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] Acquired lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.284814] env[63372]: DEBUG nova.network.neutron [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Refreshing network info cache for port 874f14b1-6cc3-4701-8a08-35940fc9bd9b {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 726.286203] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:46:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '874f14b1-6cc3-4701-8a08-35940fc9bd9b', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 726.294505] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Creating folder: Project (5cf13f5cc32c4b3ebaee47f03bb64ba7). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 726.297762] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c6621d2-c45b-43b0-bd13-a38c95cc666c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.312484] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Created folder: Project (5cf13f5cc32c4b3ebaee47f03bb64ba7) in parent group-v227230. [ 726.312484] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Creating folder: Instances. Parent ref: group-v227304. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 726.312484] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e011a2b0-4ee6-4e99-a53b-1f5d0d99df2e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.321135] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Created folder: Instances in parent group-v227304. [ 726.321135] env[63372]: DEBUG oslo.service.loopingcall [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.321135] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 726.321135] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c20c2c72-96d5-44e8-a096-3ec592544221 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.343718] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 726.343718] env[63372]: value = "task-1023889" [ 726.343718] env[63372]: _type = "Task" [ 726.343718] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.352383] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023889, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.511293] env[63372]: DEBUG nova.network.neutron [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updated VIF entry in instance network info cache for port 874f14b1-6cc3-4701-8a08-35940fc9bd9b. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 726.511675] env[63372]: DEBUG nova.network.neutron [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance_info_cache with network_info: [{"id": "874f14b1-6cc3-4701-8a08-35940fc9bd9b", "address": "fa:16:3e:57:46:8e", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.24", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap874f14b1-6c", "ovs_interfaceid": "874f14b1-6cc3-4701-8a08-35940fc9bd9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.568462] env[63372]: DEBUG nova.compute.manager [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 726.627236] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023886, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069182} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.627548] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 726.628354] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ed4b05-dbf6-48f9-b131-f731a70232d6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.644685] env[63372]: DEBUG nova.compute.manager [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 726.655579] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] f0c60559-c072-4b61-afe8-03d6c131b307/f0c60559-c072-4b61-afe8-03d6c131b307.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 726.656313] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75ad6664-03e5-4310-9fd4-4f0193a50536 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.675784] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 726.675784] env[63372]: value = "task-1023890" [ 726.675784] env[63372]: _type = "Task" [ 726.675784] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.681618] env[63372]: DEBUG nova.virt.hardware [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 726.683069] env[63372]: DEBUG nova.virt.hardware [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 726.683069] env[63372]: DEBUG nova.virt.hardware [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 726.683069] env[63372]: DEBUG nova.virt.hardware [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 726.683069] env[63372]: DEBUG nova.virt.hardware [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 726.683069] env[63372]: DEBUG nova.virt.hardware [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 726.683069] env[63372]: DEBUG nova.virt.hardware [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 726.683069] env[63372]: DEBUG nova.virt.hardware [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 726.683069] env[63372]: DEBUG nova.virt.hardware [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 726.683424] env[63372]: DEBUG nova.virt.hardware [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 726.683424] env[63372]: DEBUG nova.virt.hardware [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 726.684418] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7936a2c-c9d7-463d-a3ad-d3057f6fb8a0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.693019] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023890, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.695403] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ddac535-23df-4370-be1a-9705e726f6be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.702084] env[63372]: INFO nova.compute.manager [-] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Took 1.38 seconds to deallocate network for instance. [ 726.722567] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.103s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.726694] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.425s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.726694] env[63372]: INFO nova.compute.claims [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.749457] env[63372]: INFO nova.scheduler.client.report [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Deleted allocations for instance 276c3490-240a-499b-ba64-4b02510b169c [ 726.763871] env[63372]: DEBUG oslo_concurrency.lockutils [req-1927776b-c8c3-4ca7-aab4-ab7e04031154 req-1caf1040-ed86-4c2f-ac71-6296360008a9 service nova] Releasing lock "refresh_cache-5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.784953] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023885, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.853195] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023889, 'name': CreateVM_Task, 'duration_secs': 0.371209} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.853309] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 726.854106] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.854276] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.854596] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 726.854861] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6b228e9-ff10-4220-867b-d82e2987ba93 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.859503] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 726.859503] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52a19d9e-a61a-7636-5a1f-f7fbcc07ca19" [ 726.859503] env[63372]: _type = "Task" [ 726.859503] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.867909] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a19d9e-a61a-7636-5a1f-f7fbcc07ca19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.013944] env[63372]: DEBUG oslo_concurrency.lockutils [req-dcd6d141-3a97-4fa9-86ac-ce237cb1d262 req-9c9c0a20-0fff-465c-b8a1-3ff319ce45f5 service nova] Releasing lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.068397] env[63372]: DEBUG nova.compute.manager [req-4a05ee84-5107-46ff-bc80-8fa62d958a3a req-e3f76738-778c-423a-839b-66a8d8cb3575 service nova] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Received event network-vif-deleted-620d14ec-96a7-4912-bb91-44d2c290d431 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.090459] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.188101] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023890, 'name': ReconfigVM_Task, 'duration_secs': 0.278471} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.188454] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Reconfigured VM instance instance-0000002e to attach disk [datastore1] f0c60559-c072-4b61-afe8-03d6c131b307/f0c60559-c072-4b61-afe8-03d6c131b307.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 727.189190] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ceb80fe4-97aa-4d2f-a5d4-65d528900f49 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.195571] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 727.195571] env[63372]: value = "task-1023891" [ 727.195571] env[63372]: _type = "Task" [ 727.195571] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.203629] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023891, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.217781] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.259555] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9e8c1bbe-92bb-45c8-92cb-b125d463c3ee tempest-InstanceActionsNegativeTestJSON-1548169093 tempest-InstanceActionsNegativeTestJSON-1548169093-project-member] Lock "276c3490-240a-499b-ba64-4b02510b169c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.165s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.286118] env[63372]: DEBUG oslo_vmware.api [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023885, 'name': PowerOnVM_Task, 'duration_secs': 1.062136} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.286647] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 727.287112] env[63372]: INFO nova.compute.manager [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Took 8.48 seconds to spawn the instance on the hypervisor. [ 727.287504] env[63372]: DEBUG nova.compute.manager [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 727.288380] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b23ea22-cc08-4b81-8ad3-4346fa13052b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.370472] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a19d9e-a61a-7636-5a1f-f7fbcc07ca19, 'name': SearchDatastore_Task, 'duration_secs': 0.010131} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.370891] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.371158] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 727.371352] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.371457] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.371632] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 727.372262] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95472262-cab2-44ce-96b7-5fb48273eafb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.381173] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 727.381173] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 727.381318] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71ce4d14-715f-411f-87c3-87b75e5f9036 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.386634] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 727.386634] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fab559-3649-d1f8-c07c-e7b02ab4fa9a" [ 727.386634] env[63372]: _type = "Task" [ 727.386634] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.394439] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fab559-3649-d1f8-c07c-e7b02ab4fa9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.709904] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023891, 'name': Rename_Task, 'duration_secs': 0.156547} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.710865] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 727.711399] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f38c3d7b-883d-409c-931f-36c8b7e47e05 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.719445] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 727.719445] env[63372]: value = "task-1023892" [ 727.719445] env[63372]: _type = "Task" [ 727.719445] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.740364] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023892, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.743655] env[63372]: DEBUG nova.network.neutron [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Successfully updated port: 8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 727.753652] env[63372]: DEBUG nova.compute.manager [req-becce687-3c3b-4e4d-be84-febd772d39fd req-a817cefc-e88d-48b5-aa58-7f19131b4178 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Received event network-vif-plugged-8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.753652] env[63372]: DEBUG oslo_concurrency.lockutils [req-becce687-3c3b-4e4d-be84-febd772d39fd req-a817cefc-e88d-48b5-aa58-7f19131b4178 service nova] Acquiring lock "93a5d948-0629-4f53-a681-858d519acfa7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.753652] env[63372]: DEBUG oslo_concurrency.lockutils [req-becce687-3c3b-4e4d-be84-febd772d39fd req-a817cefc-e88d-48b5-aa58-7f19131b4178 service nova] Lock "93a5d948-0629-4f53-a681-858d519acfa7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.753846] env[63372]: DEBUG oslo_concurrency.lockutils [req-becce687-3c3b-4e4d-be84-febd772d39fd req-a817cefc-e88d-48b5-aa58-7f19131b4178 service nova] Lock "93a5d948-0629-4f53-a681-858d519acfa7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.753983] env[63372]: DEBUG nova.compute.manager [req-becce687-3c3b-4e4d-be84-febd772d39fd req-a817cefc-e88d-48b5-aa58-7f19131b4178 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] No waiting events found dispatching network-vif-plugged-8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 727.754314] env[63372]: WARNING nova.compute.manager [req-becce687-3c3b-4e4d-be84-febd772d39fd req-a817cefc-e88d-48b5-aa58-7f19131b4178 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Received unexpected event network-vif-plugged-8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f for instance with vm_state building and task_state spawning. [ 727.815603] env[63372]: INFO nova.compute.manager [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Took 37.26 seconds to build instance. [ 727.904239] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fab559-3649-d1f8-c07c-e7b02ab4fa9a, 'name': SearchDatastore_Task, 'duration_secs': 0.008618} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.905147] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0618acfb-a50f-4edf-8e45-3307c4fc50fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.915433] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 727.915433] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5270c21c-a1f5-81db-4d37-e0b1ad438481" [ 727.915433] env[63372]: _type = "Task" [ 727.915433] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.923480] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5270c21c-a1f5-81db-4d37-e0b1ad438481, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.173804] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be5bdd3-bd92-4cec-882c-91ebd692eae2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.182194] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c07d734-7951-499b-844d-b74d44928eac {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.217041] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c25235-5f6b-443a-89bb-b581ebee3832 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.227130] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce502ed-238f-496a-bb9c-16a57fd0ef0c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.236725] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023892, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.245012] env[63372]: DEBUG nova.compute.provider_tree [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.246649] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.246872] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.246872] env[63372]: DEBUG nova.network.neutron [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 728.317999] env[63372]: DEBUG oslo_concurrency.lockutils [None req-15affb45-adac-434d-aa4a-bc0db2edc74a tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "b182294d-2de8-4189-af7f-3e2d2c604a8b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.126s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.425122] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5270c21c-a1f5-81db-4d37-e0b1ad438481, 'name': SearchDatastore_Task, 'duration_secs': 0.009845} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.425122] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.426024] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 8283b736-ad02-4082-97b7-561bd5c5da93/8283b736-ad02-4082-97b7-561bd5c5da93.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 728.426024] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-053b2e7c-e0c3-4df5-b4f5-4a6afc435086 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.432030] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 728.432030] env[63372]: value = "task-1023893" [ 728.432030] env[63372]: _type = "Task" [ 728.432030] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.439678] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1023893, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.738763] env[63372]: DEBUG oslo_vmware.api [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023892, 'name': PowerOnVM_Task, 'duration_secs': 0.569068} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.738763] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 728.738763] env[63372]: INFO nova.compute.manager [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Took 7.31 seconds to spawn the instance on the hypervisor. [ 728.741224] env[63372]: DEBUG nova.compute.manager [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 728.741224] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5610dd7e-1833-47c5-b69a-8c1905d0e4cd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.752403] env[63372]: DEBUG nova.scheduler.client.report [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.788205] env[63372]: DEBUG nova.network.neutron [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.824226] env[63372]: DEBUG nova.compute.manager [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 728.949083] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1023893, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.960868] env[63372]: DEBUG nova.compute.manager [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 728.961733] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8944a7d7-513b-49ac-8194-143cd1f7debf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.000561] env[63372]: DEBUG nova.network.neutron [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance_info_cache with network_info: [{"id": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "address": "fa:16:3e:3f:99:77", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c4c74ba-ad", "ovs_interfaceid": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.257344] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.257883] env[63372]: DEBUG nova.compute.manager [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 729.266169] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 23.665s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.266169] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.266169] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63372) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 729.266169] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.259s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.266169] env[63372]: DEBUG nova.objects.instance [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Lazy-loading 'resources' on Instance uuid 44fba254-7dec-4458-9ae3-fdbbe4895de5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 729.271518] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44ec744-462d-4781-9c08-6f50f992c230 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.280902] env[63372]: INFO nova.compute.manager [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Took 36.63 seconds to build instance. [ 729.288490] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b847ab4d-85de-4bda-a6e0-07f5ef1a281d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.313849] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efdf550f-b91e-4f2b-8b09-97679c627cd9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.324179] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6749fe-fae5-4252-a6a2-0aa1c4296a93 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.366071] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180529MB free_disk=185GB free_vcpus=48 pci_devices=None {{(pid=63372) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 729.366211] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.386016] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.449381] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1023893, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.472948] env[63372]: INFO nova.compute.manager [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] instance snapshotting [ 729.476234] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53251215-43b9-4264-ab81-eeffb6770409 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.498847] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a8178a-7064-4d9d-ad11-d28b36b2c6cd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.503835] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.503961] env[63372]: DEBUG nova.compute.manager [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Instance network_info: |[{"id": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "address": "fa:16:3e:3f:99:77", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c4c74ba-ad", "ovs_interfaceid": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 729.507124] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3f:99:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d5d0e0d-cdec-474a-a891-a9ceff15a8b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 729.513741] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Creating folder: Project (5e46b6acae4a4b71bc390bd513ddd2f1). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 729.516331] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a70199e3-ba97-4534-8d34-aff860d0b2dc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.525816] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Created folder: Project (5e46b6acae4a4b71bc390bd513ddd2f1) in parent group-v227230. [ 729.526015] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Creating folder: Instances. Parent ref: group-v227307. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 729.526259] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-14976154-2739-466d-aff9-9865d23464a6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.535710] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Created folder: Instances in parent group-v227307. [ 729.536050] env[63372]: DEBUG oslo.service.loopingcall [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.536173] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 729.536349] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-08d26c90-38d6-4454-89f4-49925a410513 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.554838] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 729.554838] env[63372]: value = "task-1023896" [ 729.554838] env[63372]: _type = "Task" [ 729.554838] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.563625] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023896, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.766976] env[63372]: DEBUG nova.compute.utils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 729.769210] env[63372]: DEBUG nova.compute.manager [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 729.769210] env[63372]: DEBUG nova.network.neutron [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 729.784943] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2da82e77-a102-4ea1-ba38-ad4a78c33c7c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "f0c60559-c072-4b61-afe8-03d6c131b307" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.381s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.788162] env[63372]: DEBUG nova.compute.manager [req-27ac57c7-19a1-4508-bb28-6a47dc409e1c req-3148f976-a9f1-4993-b8ff-8e1fc17bd4f0 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Received event network-changed-8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.788786] env[63372]: DEBUG nova.compute.manager [req-27ac57c7-19a1-4508-bb28-6a47dc409e1c req-3148f976-a9f1-4993-b8ff-8e1fc17bd4f0 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Refreshing instance network info cache due to event network-changed-8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 729.788786] env[63372]: DEBUG oslo_concurrency.lockutils [req-27ac57c7-19a1-4508-bb28-6a47dc409e1c req-3148f976-a9f1-4993-b8ff-8e1fc17bd4f0 service nova] Acquiring lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.788786] env[63372]: DEBUG oslo_concurrency.lockutils [req-27ac57c7-19a1-4508-bb28-6a47dc409e1c req-3148f976-a9f1-4993-b8ff-8e1fc17bd4f0 service nova] Acquired lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.788989] env[63372]: DEBUG nova.network.neutron [req-27ac57c7-19a1-4508-bb28-6a47dc409e1c req-3148f976-a9f1-4993-b8ff-8e1fc17bd4f0 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Refreshing network info cache for port 8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 729.817099] env[63372]: DEBUG nova.policy [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03d36cf8ed3d4c6a8dc9053a38e08d29', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a1f1ccc886484781af500807c878239b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 729.952797] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1023893, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.070234} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.953320] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 8283b736-ad02-4082-97b7-561bd5c5da93/8283b736-ad02-4082-97b7-561bd5c5da93.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 729.953589] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 729.953883] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea7abf6e-9101-43f8-814a-1321753f4ab8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.963703] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 729.963703] env[63372]: value = "task-1023897" [ 729.963703] env[63372]: _type = "Task" [ 729.963703] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.976733] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1023897, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.022073] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 730.022073] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-65b5cfe1-4e74-40c1-8ea5-97db602b42c0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.029769] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 730.029769] env[63372]: value = "task-1023898" [ 730.029769] env[63372]: _type = "Task" [ 730.029769] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.042028] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023898, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.065618] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023896, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.203165] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0998be18-bae8-492b-a801-2d5baee9ce76 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.206036] env[63372]: DEBUG nova.network.neutron [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Successfully created port: 6e151063-238b-4403-9365-be674f0638f5 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.213560] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630d6cbc-c8e5-47af-bada-2d6d9600e42b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.246820] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-480ec816-9a65-4b73-a956-1de13e1cbcd0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.255042] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be258c0e-1649-439c-be07-dacddb126e05 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.270385] env[63372]: DEBUG nova.compute.provider_tree [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.271725] env[63372]: DEBUG nova.compute.manager [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 730.291108] env[63372]: DEBUG nova.compute.manager [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 730.476152] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1023897, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108733} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.476496] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 730.477285] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72063c27-e4c2-43a0-9a47-744f5161c50f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.501532] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] 8283b736-ad02-4082-97b7-561bd5c5da93/8283b736-ad02-4082-97b7-561bd5c5da93.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 730.501832] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d13f15d-43d5-4e5d-af61-783eed71c6ab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.519266] env[63372]: DEBUG nova.network.neutron [req-27ac57c7-19a1-4508-bb28-6a47dc409e1c req-3148f976-a9f1-4993-b8ff-8e1fc17bd4f0 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updated VIF entry in instance network info cache for port 8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 730.522085] env[63372]: DEBUG nova.network.neutron [req-27ac57c7-19a1-4508-bb28-6a47dc409e1c req-3148f976-a9f1-4993-b8ff-8e1fc17bd4f0 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance_info_cache with network_info: [{"id": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "address": "fa:16:3e:3f:99:77", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c4c74ba-ad", "ovs_interfaceid": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.527098] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 730.527098] env[63372]: value = "task-1023899" [ 730.527098] env[63372]: _type = "Task" [ 730.527098] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.540816] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1023899, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.546010] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023898, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.570780] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023896, 'name': CreateVM_Task, 'duration_secs': 0.516092} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.570920] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 730.571556] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.571698] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.572018] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 730.572668] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3be2ac3a-191b-4373-ba04-1a260c5574fe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.579444] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 730.579444] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52ba365a-eca4-c8f6-3dc5-bd154803191f" [ 730.579444] env[63372]: _type = "Task" [ 730.579444] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.587990] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ba365a-eca4-c8f6-3dc5-bd154803191f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.775401] env[63372]: DEBUG nova.scheduler.client.report [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.821212] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.022460] env[63372]: DEBUG oslo_concurrency.lockutils [req-27ac57c7-19a1-4508-bb28-6a47dc409e1c req-3148f976-a9f1-4993-b8ff-8e1fc17bd4f0 service nova] Releasing lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.040584] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1023899, 'name': ReconfigVM_Task, 'duration_secs': 0.503471} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.040883] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Reconfigured VM instance instance-0000002f to attach disk [datastore2] 8283b736-ad02-4082-97b7-561bd5c5da93/8283b736-ad02-4082-97b7-561bd5c5da93.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 731.042046] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9662664a-32d5-4707-a446-e953f707c217 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.046544] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023898, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.050857] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 731.050857] env[63372]: value = "task-1023900" [ 731.050857] env[63372]: _type = "Task" [ 731.050857] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.061842] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1023900, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.089034] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ba365a-eca4-c8f6-3dc5-bd154803191f, 'name': SearchDatastore_Task, 'duration_secs': 0.032424} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.089366] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.089608] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 731.089917] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.090037] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.090159] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 731.090436] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-275e7ae9-87d6-479d-bae9-2dc4ecc9b6b2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.098820] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 731.099028] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 731.099777] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63e0aec3-2daa-4096-bee8-08ca15e1be9b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.104962] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 731.104962] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e16889-6ae7-113f-59e9-88e65cce7bb4" [ 731.104962] env[63372]: _type = "Task" [ 731.104962] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.112668] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e16889-6ae7-113f-59e9-88e65cce7bb4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.199302] env[63372]: DEBUG nova.compute.manager [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 731.199690] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f36becc-7f8d-4e43-a68a-4b5b44a047ee {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.285680] env[63372]: DEBUG nova.compute.manager [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 731.288368] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.022s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.290654] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.245s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.292200] env[63372]: INFO nova.compute.claims [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.317707] env[63372]: INFO nova.scheduler.client.report [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Deleted allocations for instance 44fba254-7dec-4458-9ae3-fdbbe4895de5 [ 731.324092] env[63372]: DEBUG nova.virt.hardware [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 731.324574] env[63372]: DEBUG nova.virt.hardware [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 731.324574] env[63372]: DEBUG nova.virt.hardware [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 731.324756] env[63372]: DEBUG nova.virt.hardware [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 731.325429] env[63372]: DEBUG nova.virt.hardware [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 731.325429] env[63372]: DEBUG nova.virt.hardware [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 731.325429] env[63372]: DEBUG nova.virt.hardware [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 731.325429] env[63372]: DEBUG nova.virt.hardware [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 731.325629] env[63372]: DEBUG nova.virt.hardware [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 731.325749] env[63372]: DEBUG nova.virt.hardware [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 731.325959] env[63372]: DEBUG nova.virt.hardware [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 731.327174] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d6eef8-2f55-4e4f-abeb-6abc44010a1f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.339965] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bcebc45-668c-4d02-a1d8-93ed8547898a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.544715] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023898, 'name': CreateSnapshot_Task, 'duration_secs': 1.241795} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.544958] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 731.545674] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed56368f-a50f-40cd-9808-2702eefab86b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.561548] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1023900, 'name': Rename_Task, 'duration_secs': 0.169121} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.561778] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 731.561984] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b948f0e-bfe0-4b83-ac8d-1bfb0749d271 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.568103] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 731.568103] env[63372]: value = "task-1023901" [ 731.568103] env[63372]: _type = "Task" [ 731.568103] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.575341] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1023901, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.615280] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e16889-6ae7-113f-59e9-88e65cce7bb4, 'name': SearchDatastore_Task, 'duration_secs': 0.010198} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.616051] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d2b8808-f238-47c9-acc8-231db0be0ba3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.622087] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 731.622087] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f76b0c-e952-0be1-d610-3a055372629f" [ 731.622087] env[63372]: _type = "Task" [ 731.622087] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.629911] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f76b0c-e952-0be1-d610-3a055372629f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.712301] env[63372]: INFO nova.compute.manager [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] instance snapshotting [ 731.715635] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9c60dc-0d51-41c7-803b-eb215f956bc6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.737460] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae471332-b824-4e67-a81c-48f4bc736d06 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.837232] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d333a0e8-cb61-4d97-be69-f97fb18df5e5 tempest-ServersAaction247Test-1502560618 tempest-ServersAaction247Test-1502560618-project-member] Lock "44fba254-7dec-4458-9ae3-fdbbe4895de5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.604s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.066103] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 732.066409] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ebcd91e4-a8a0-4fcc-95fb-c8bc7022866f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.079018] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1023901, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.080324] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 732.080324] env[63372]: value = "task-1023902" [ 732.080324] env[63372]: _type = "Task" [ 732.080324] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.092479] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023902, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.121441] env[63372]: DEBUG nova.compute.manager [req-e34aa6a5-9d37-49b6-8785-7985e9388881 req-2f82c8bf-2f5e-4404-9058-4fd0933cf0ae service nova] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Received event network-vif-plugged-6e151063-238b-4403-9365-be674f0638f5 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 732.121690] env[63372]: DEBUG oslo_concurrency.lockutils [req-e34aa6a5-9d37-49b6-8785-7985e9388881 req-2f82c8bf-2f5e-4404-9058-4fd0933cf0ae service nova] Acquiring lock "f43555ef-b517-4b7d-9d2e-4787d40e201b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.121910] env[63372]: DEBUG oslo_concurrency.lockutils [req-e34aa6a5-9d37-49b6-8785-7985e9388881 req-2f82c8bf-2f5e-4404-9058-4fd0933cf0ae service nova] Lock "f43555ef-b517-4b7d-9d2e-4787d40e201b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.122151] env[63372]: DEBUG oslo_concurrency.lockutils [req-e34aa6a5-9d37-49b6-8785-7985e9388881 req-2f82c8bf-2f5e-4404-9058-4fd0933cf0ae service nova] Lock "f43555ef-b517-4b7d-9d2e-4787d40e201b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.122250] env[63372]: DEBUG nova.compute.manager [req-e34aa6a5-9d37-49b6-8785-7985e9388881 req-2f82c8bf-2f5e-4404-9058-4fd0933cf0ae service nova] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] No waiting events found dispatching network-vif-plugged-6e151063-238b-4403-9365-be674f0638f5 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 732.122461] env[63372]: WARNING nova.compute.manager [req-e34aa6a5-9d37-49b6-8785-7985e9388881 req-2f82c8bf-2f5e-4404-9058-4fd0933cf0ae service nova] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Received unexpected event network-vif-plugged-6e151063-238b-4403-9365-be674f0638f5 for instance with vm_state building and task_state spawning. [ 732.135060] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f76b0c-e952-0be1-d610-3a055372629f, 'name': SearchDatastore_Task, 'duration_secs': 0.019899} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.135307] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.135550] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 93a5d948-0629-4f53-a681-858d519acfa7/93a5d948-0629-4f53-a681-858d519acfa7.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 732.135800] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4eccdcce-896e-4911-96b8-e7e3ea523213 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.142415] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 732.142415] env[63372]: value = "task-1023903" [ 732.142415] env[63372]: _type = "Task" [ 732.142415] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.151944] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1023903, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.224437] env[63372]: DEBUG nova.network.neutron [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Successfully updated port: 6e151063-238b-4403-9365-be674f0638f5 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 732.250184] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 732.251743] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-dafdb519-6cbd-49b9-8b64-08a9d05f9865 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.261372] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 732.261372] env[63372]: value = "task-1023904" [ 732.261372] env[63372]: _type = "Task" [ 732.261372] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.270598] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023904, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.583790] env[63372]: DEBUG oslo_vmware.api [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1023901, 'name': PowerOnVM_Task, 'duration_secs': 0.888021} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.588413] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 732.588800] env[63372]: INFO nova.compute.manager [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Took 8.52 seconds to spawn the instance on the hypervisor. [ 732.589053] env[63372]: DEBUG nova.compute.manager [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 732.590203] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d17ea4-ef9a-4fd4-b188-a9bfb87c5dfa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.605841] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023902, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.656248] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1023903, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.720782] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c690c18e-b3c3-475c-a6f5-c90116be2c0f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.728066] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3468e288-6329-4b57-99ba-702b1633b598 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.731793] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Acquiring lock "refresh_cache-f43555ef-b517-4b7d-9d2e-4787d40e201b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.731997] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Acquired lock "refresh_cache-f43555ef-b517-4b7d-9d2e-4787d40e201b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.732086] env[63372]: DEBUG nova.network.neutron [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 732.759209] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb38a9bb-bf24-4e21-8293-8f59fbaeb2e2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.774317] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038b2296-24d1-429a-aeb0-95121f653752 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.778276] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023904, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.788792] env[63372]: DEBUG nova.compute.provider_tree [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.093535] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023902, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.117157] env[63372]: INFO nova.compute.manager [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Took 34.80 seconds to build instance. [ 733.155286] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1023903, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.536858} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.155286] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 93a5d948-0629-4f53-a681-858d519acfa7/93a5d948-0629-4f53-a681-858d519acfa7.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 733.155286] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 733.155449] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ac42247e-c416-4354-a606-361b577da18a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.162924] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 733.162924] env[63372]: value = "task-1023905" [ 733.162924] env[63372]: _type = "Task" [ 733.162924] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.172330] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1023905, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.272545] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023904, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.292794] env[63372]: DEBUG nova.scheduler.client.report [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.305265] env[63372]: DEBUG nova.network.neutron [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.559146] env[63372]: DEBUG nova.network.neutron [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Updating instance_info_cache with network_info: [{"id": "6e151063-238b-4403-9365-be674f0638f5", "address": "fa:16:3e:90:77:35", "network": {"id": "9d91d339-76e3-4cd6-b317-7bdef7701294", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-458333239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a1f1ccc886484781af500807c878239b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d12aff80-9d1b-4a67-a470-9c0148b443e3", "external-id": "nsx-vlan-transportzone-784", "segmentation_id": 784, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e151063-23", "ovs_interfaceid": "6e151063-238b-4403-9365-be674f0638f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.597421] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023902, 'name': CloneVM_Task} progress is 95%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.609990] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "3f66eda6-f5e9-4527-9711-849a01702580" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.611650] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "3f66eda6-f5e9-4527-9711-849a01702580" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.620247] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d4ce99f-204a-491b-9586-48fc4190fb7e tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "8283b736-ad02-4082-97b7-561bd5c5da93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.750s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.674478] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1023905, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069587} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.674478] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 733.674922] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55019972-b863-47fc-b3ee-ea6ae50ff883 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.701237] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 93a5d948-0629-4f53-a681-858d519acfa7/93a5d948-0629-4f53-a681-858d519acfa7.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 733.702022] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53fdbe9d-4cc7-42dc-8cc0-ed3129f98559 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.726931] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 733.726931] env[63372]: value = "task-1023906" [ 733.726931] env[63372]: _type = "Task" [ 733.726931] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.737935] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1023906, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.773658] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023904, 'name': CreateSnapshot_Task, 'duration_secs': 1.091527} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.775214] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 733.778272] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1dd48cf-1b57-45d9-97aa-485502cc4707 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.781228] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "f6675697-1529-46be-b28a-398ff3060d18" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.781432] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "f6675697-1529-46be-b28a-398ff3060d18" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.798274] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.508s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.799124] env[63372]: DEBUG nova.compute.manager [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.803380] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.677s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.803738] env[63372]: DEBUG nova.objects.instance [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Lazy-loading 'resources' on Instance uuid 85883c62-6d5e-4fa6-b4f0-afac984dc6bb {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 734.063837] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Releasing lock "refresh_cache-f43555ef-b517-4b7d-9d2e-4787d40e201b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.064173] env[63372]: DEBUG nova.compute.manager [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Instance network_info: |[{"id": "6e151063-238b-4403-9365-be674f0638f5", "address": "fa:16:3e:90:77:35", "network": {"id": "9d91d339-76e3-4cd6-b317-7bdef7701294", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-458333239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a1f1ccc886484781af500807c878239b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d12aff80-9d1b-4a67-a470-9c0148b443e3", "external-id": "nsx-vlan-transportzone-784", "segmentation_id": 784, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e151063-23", "ovs_interfaceid": "6e151063-238b-4403-9365-be674f0638f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 734.064599] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:77:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd12aff80-9d1b-4a67-a470-9c0148b443e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6e151063-238b-4403-9365-be674f0638f5', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 734.073380] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Creating folder: Project (a1f1ccc886484781af500807c878239b). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 734.073726] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-58c45350-fd73-4132-b505-5d4ae6a55056 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.084517] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Created folder: Project (a1f1ccc886484781af500807c878239b) in parent group-v227230. [ 734.084700] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Creating folder: Instances. Parent ref: group-v227313. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 734.084948] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b0238873-81b7-414c-9e6f-470e0c2b472d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.095599] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023902, 'name': CloneVM_Task, 'duration_secs': 1.860299} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.097039] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Created linked-clone VM from snapshot [ 734.097325] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Created folder: Instances in parent group-v227313. [ 734.097551] env[63372]: DEBUG oslo.service.loopingcall [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.098231] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5efdfbe8-5c19-48ed-a5e3-610a49a4739d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.100690] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 734.100894] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cff39d41-3d4d-4829-a15f-3941f664a96f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.120938] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Uploading image 18c11831-fcc8-4c06-9487-9dee2888b9f7 {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 734.124212] env[63372]: DEBUG nova.compute.manager [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 734.126944] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 734.126944] env[63372]: value = "task-1023909" [ 734.126944] env[63372]: _type = "Task" [ 734.126944] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.134962] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023909, 'name': CreateVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.136621] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 734.136922] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7486fd2c-1f21-4a42-9c16-e5383ea3f4a3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.144018] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 734.144018] env[63372]: value = "task-1023910" [ 734.144018] env[63372]: _type = "Task" [ 734.144018] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.154651] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023910, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.157285] env[63372]: DEBUG nova.compute.manager [req-6dcbb8b0-5ece-4fa5-8095-08eaa55dbd40 req-e75fd2f2-2f94-41dd-b7f2-9fb250fc2359 service nova] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Received event network-changed-6e151063-238b-4403-9365-be674f0638f5 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 734.157463] env[63372]: DEBUG nova.compute.manager [req-6dcbb8b0-5ece-4fa5-8095-08eaa55dbd40 req-e75fd2f2-2f94-41dd-b7f2-9fb250fc2359 service nova] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Refreshing instance network info cache due to event network-changed-6e151063-238b-4403-9365-be674f0638f5. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 734.157678] env[63372]: DEBUG oslo_concurrency.lockutils [req-6dcbb8b0-5ece-4fa5-8095-08eaa55dbd40 req-e75fd2f2-2f94-41dd-b7f2-9fb250fc2359 service nova] Acquiring lock "refresh_cache-f43555ef-b517-4b7d-9d2e-4787d40e201b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.157847] env[63372]: DEBUG oslo_concurrency.lockutils [req-6dcbb8b0-5ece-4fa5-8095-08eaa55dbd40 req-e75fd2f2-2f94-41dd-b7f2-9fb250fc2359 service nova] Acquired lock "refresh_cache-f43555ef-b517-4b7d-9d2e-4787d40e201b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.158214] env[63372]: DEBUG nova.network.neutron [req-6dcbb8b0-5ece-4fa5-8095-08eaa55dbd40 req-e75fd2f2-2f94-41dd-b7f2-9fb250fc2359 service nova] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Refreshing network info cache for port 6e151063-238b-4403-9365-be674f0638f5 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 734.249040] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1023906, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.298783] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 734.301909] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-84b13c5b-49e6-47e9-a533-167eb6a2b1b2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.308623] env[63372]: DEBUG nova.compute.utils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 734.314023] env[63372]: DEBUG nova.compute.manager [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 734.314023] env[63372]: DEBUG nova.network.neutron [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 734.320851] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 734.320851] env[63372]: value = "task-1023911" [ 734.320851] env[63372]: _type = "Task" [ 734.320851] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.145907] env[63372]: DEBUG nova.compute.manager [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 735.149251] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023911, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.149514] env[63372]: WARNING oslo_vmware.common.loopingcall [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] task run outlasted interval by 0.328364 sec [ 735.151359] env[63372]: DEBUG nova.policy [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73c19d06a5a4433fb71d9e5179399a57', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9946a38bddee4f7cb2c6551d76f0a1c0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 735.183664] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023910, 'name': Destroy_Task, 'duration_secs': 0.42507} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.193374] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Destroyed the VM [ 735.193885] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 735.195173] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023911, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.195173] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023909, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.195173] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1023906, 'name': ReconfigVM_Task, 'duration_secs': 0.650236} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.195929] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.196192] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f569ad9a-94dc-463d-b193-e2fdbc785a9a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.198279] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 93a5d948-0629-4f53-a681-858d519acfa7/93a5d948-0629-4f53-a681-858d519acfa7.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 735.203708] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dde32613-507b-4e9c-a0e8-5441e84b5c1b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.212016] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 735.212016] env[63372]: value = "task-1023912" [ 735.212016] env[63372]: _type = "Task" [ 735.212016] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.212016] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 735.212016] env[63372]: value = "task-1023913" [ 735.212016] env[63372]: _type = "Task" [ 735.212016] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.233019] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023912, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.237890] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1023913, 'name': Rename_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.441874] env[63372]: DEBUG nova.network.neutron [req-6dcbb8b0-5ece-4fa5-8095-08eaa55dbd40 req-e75fd2f2-2f94-41dd-b7f2-9fb250fc2359 service nova] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Updated VIF entry in instance network info cache for port 6e151063-238b-4403-9365-be674f0638f5. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 735.441874] env[63372]: DEBUG nova.network.neutron [req-6dcbb8b0-5ece-4fa5-8095-08eaa55dbd40 req-e75fd2f2-2f94-41dd-b7f2-9fb250fc2359 service nova] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Updating instance_info_cache with network_info: [{"id": "6e151063-238b-4403-9365-be674f0638f5", "address": "fa:16:3e:90:77:35", "network": {"id": "9d91d339-76e3-4cd6-b317-7bdef7701294", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-458333239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a1f1ccc886484781af500807c878239b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d12aff80-9d1b-4a67-a470-9c0148b443e3", "external-id": "nsx-vlan-transportzone-784", "segmentation_id": 784, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e151063-23", "ovs_interfaceid": "6e151063-238b-4403-9365-be674f0638f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.620029] env[63372]: DEBUG nova.network.neutron [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Successfully created port: 6c9c3162-8734-4911-a9d1-9f3f9214ea93 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 735.637070] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2226f73-5fd8-4d05-ba18-3506164b499d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.652018] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6beebb0c-f154-42cf-9876-faa15cf93af8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.655277] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023909, 'name': CreateVM_Task, 'duration_secs': 1.408597} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.655460] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 735.656463] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.656652] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.656962] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 735.657227] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-471cd630-b653-41e1-9748-afc1454035ad {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.700101] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e867c6-2e3e-4a6e-8e04-601fb4e92168 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.704771] env[63372]: DEBUG nova.compute.manager [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Stashing vm_state: active {{(pid=63372) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 735.709303] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 735.709303] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]524b701d-e9e7-ecfc-244d-434e51f7b62b" [ 735.709303] env[63372]: _type = "Task" [ 735.709303] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.721475] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023911, 'name': CloneVM_Task} progress is 95%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.728722] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19aab525-d0ce-4afa-98a3-120e7e5906e8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.736426] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524b701d-e9e7-ecfc-244d-434e51f7b62b, 'name': SearchDatastore_Task, 'duration_secs': 0.013899} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.737611] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.737886] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 735.738131] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.738371] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.738446] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 735.744612] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b9b412b5-b430-4114-9538-3bb16960ef76 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.746478] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1023913, 'name': Rename_Task, 'duration_secs': 0.400598} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.746747] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023912, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.757024] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 735.757024] env[63372]: DEBUG nova.compute.provider_tree [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.761023] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c596bdbe-7953-44d7-a0b1-1f9330713d13 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.765731] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 735.765731] env[63372]: value = "task-1023914" [ 735.765731] env[63372]: _type = "Task" [ 735.765731] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.766280] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 735.766280] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 735.770118] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d33a36bd-d838-4739-9443-b1860a85530e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.782110] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1023914, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.782110] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 735.782110] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5268425f-33bd-59ff-e78c-a1845feb176d" [ 735.782110] env[63372]: _type = "Task" [ 735.782110] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.787659] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5268425f-33bd-59ff-e78c-a1845feb176d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.943697] env[63372]: DEBUG oslo_concurrency.lockutils [req-6dcbb8b0-5ece-4fa5-8095-08eaa55dbd40 req-e75fd2f2-2f94-41dd-b7f2-9fb250fc2359 service nova] Releasing lock "refresh_cache-f43555ef-b517-4b7d-9d2e-4787d40e201b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.199089] env[63372]: DEBUG nova.compute.manager [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 736.201112] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023911, 'name': CloneVM_Task, 'duration_secs': 1.629354} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.201548] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Created linked-clone VM from snapshot [ 736.202333] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de8e362-7da5-49c9-be7a-a3c373c0ba0b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.209377] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Uploading image 184e52b2-673c-4afd-b14d-dcd92d374426 {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 736.223077] env[63372]: DEBUG nova.virt.hardware [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 736.223347] env[63372]: DEBUG nova.virt.hardware [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 736.223504] env[63372]: DEBUG nova.virt.hardware [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 736.223718] env[63372]: DEBUG nova.virt.hardware [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 736.223867] env[63372]: DEBUG nova.virt.hardware [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 736.224444] env[63372]: DEBUG nova.virt.hardware [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 736.224444] env[63372]: DEBUG nova.virt.hardware [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 736.224444] env[63372]: DEBUG nova.virt.hardware [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 736.224584] env[63372]: DEBUG nova.virt.hardware [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 736.224664] env[63372]: DEBUG nova.virt.hardware [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 736.224831] env[63372]: DEBUG nova.virt.hardware [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 736.225737] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-687c5bd5-aa3d-462c-96f1-9a5b732be0b5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.231633] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.240301] env[63372]: DEBUG oslo_vmware.api [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023912, 'name': RemoveSnapshot_Task, 'duration_secs': 0.963009} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.242498] env[63372]: DEBUG oslo_vmware.rw_handles [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 736.242498] env[63372]: value = "vm-227316" [ 736.242498] env[63372]: _type = "VirtualMachine" [ 736.242498] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 736.244643] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 736.247029] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c7963f68-ec53-4558-88a2-e294b526837e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.250027] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4857ca94-dbe1-45ab-a497-4be5e434bb63 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.264639] env[63372]: DEBUG nova.scheduler.client.report [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 736.270044] env[63372]: DEBUG oslo_vmware.rw_handles [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lease: (returnval){ [ 736.270044] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520007f5-f193-aa75-037b-d8f49c043d6a" [ 736.270044] env[63372]: _type = "HttpNfcLease" [ 736.270044] env[63372]: } obtained for exporting VM: (result){ [ 736.270044] env[63372]: value = "vm-227316" [ 736.270044] env[63372]: _type = "VirtualMachine" [ 736.270044] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 736.270044] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the lease: (returnval){ [ 736.270044] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520007f5-f193-aa75-037b-d8f49c043d6a" [ 736.270044] env[63372]: _type = "HttpNfcLease" [ 736.270044] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 736.281278] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1023914, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.285582] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 736.285582] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520007f5-f193-aa75-037b-d8f49c043d6a" [ 736.285582] env[63372]: _type = "HttpNfcLease" [ 736.285582] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 736.292328] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5268425f-33bd-59ff-e78c-a1845feb176d, 'name': SearchDatastore_Task, 'duration_secs': 0.011202} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.293129] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e170a1d-2015-4981-b416-f7117bfea071 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.298295] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 736.298295] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]527edf66-4517-e6b8-36de-7cd0952c9402" [ 736.298295] env[63372]: _type = "Task" [ 736.298295] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.306303] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527edf66-4517-e6b8-36de-7cd0952c9402, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.752319] env[63372]: WARNING nova.compute.manager [None req-6219551b-062d-4607-ac3d-63746c0bef4b tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Image not found during snapshot: nova.exception.ImageNotFound: Image 18c11831-fcc8-4c06-9487-9dee2888b9f7 could not be found. [ 736.774148] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.971s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.777564] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.938s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.780027] env[63372]: INFO nova.compute.claims [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.791076] env[63372]: DEBUG oslo_vmware.api [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1023914, 'name': PowerOnVM_Task, 'duration_secs': 0.898042} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.792925] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 736.793205] env[63372]: INFO nova.compute.manager [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Took 10.15 seconds to spawn the instance on the hypervisor. [ 736.793429] env[63372]: DEBUG nova.compute.manager [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 736.793721] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 736.793721] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520007f5-f193-aa75-037b-d8f49c043d6a" [ 736.793721] env[63372]: _type = "HttpNfcLease" [ 736.793721] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 736.794702] env[63372]: INFO nova.scheduler.client.report [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Deleted allocations for instance 85883c62-6d5e-4fa6-b4f0-afac984dc6bb [ 736.797107] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba109243-f2c7-4593-b674-785dd9f3188f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.799754] env[63372]: DEBUG oslo_vmware.rw_handles [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 736.799754] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520007f5-f193-aa75-037b-d8f49c043d6a" [ 736.799754] env[63372]: _type = "HttpNfcLease" [ 736.799754] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 736.803654] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18408fea-7ee7-4ea2-9872-e3685c8044ae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.824395] env[63372]: DEBUG oslo_vmware.rw_handles [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5206ac3d-20d3-e247-b36b-ab82c883fb3a/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 736.824633] env[63372]: DEBUG oslo_vmware.rw_handles [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5206ac3d-20d3-e247-b36b-ab82c883fb3a/disk-0.vmdk for reading. {{(pid=63372) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 736.825936] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527edf66-4517-e6b8-36de-7cd0952c9402, 'name': SearchDatastore_Task, 'duration_secs': 0.016381} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.827055] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.827328] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] f43555ef-b517-4b7d-9d2e-4787d40e201b/f43555ef-b517-4b7d-9d2e-4787d40e201b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 736.828458] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-06960445-7b96-4617-9cb8-7d65174daf36 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.893020] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 736.893020] env[63372]: value = "task-1023916" [ 736.893020] env[63372]: _type = "Task" [ 736.893020] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.903538] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023916, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.945239] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-75048c36-34f3-4a85-abcb-a30d195b0032 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.310846] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8ec546d6-564a-46c2-abc7-ad9fca4b3099 tempest-ServersNegativeTestMultiTenantJSON-1166880086 tempest-ServersNegativeTestMultiTenantJSON-1166880086-project-member] Lock "85883c62-6d5e-4fa6-b4f0-afac984dc6bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.109s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.343024] env[63372]: INFO nova.compute.manager [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Took 37.45 seconds to build instance. [ 737.402018] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023916, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.845628] env[63372]: DEBUG oslo_concurrency.lockutils [None req-587c757c-4ce6-4628-9a2f-cf1dc983f4c1 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "93a5d948-0629-4f53-a681-858d519acfa7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.286s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.913092] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023916, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.696008} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.913484] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] f43555ef-b517-4b7d-9d2e-4787d40e201b/f43555ef-b517-4b7d-9d2e-4787d40e201b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 737.913732] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 737.914082] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c99bf37e-daa7-4a80-a0ef-86bee29a0b29 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.923418] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 737.923418] env[63372]: value = "task-1023917" [ 737.923418] env[63372]: _type = "Task" [ 737.923418] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.926065] env[63372]: DEBUG nova.compute.manager [req-15d8a8b6-244e-41af-a599-49e4d17afaef req-7d265f98-2acc-493a-9755-ff58e3225c68 service nova] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Received event network-vif-plugged-6c9c3162-8734-4911-a9d1-9f3f9214ea93 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.926384] env[63372]: DEBUG oslo_concurrency.lockutils [req-15d8a8b6-244e-41af-a599-49e4d17afaef req-7d265f98-2acc-493a-9755-ff58e3225c68 service nova] Acquiring lock "7a75a51d-b804-453b-ba7b-fda6c4931802-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.931616] env[63372]: DEBUG oslo_concurrency.lockutils [req-15d8a8b6-244e-41af-a599-49e4d17afaef req-7d265f98-2acc-493a-9755-ff58e3225c68 service nova] Lock "7a75a51d-b804-453b-ba7b-fda6c4931802-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.931616] env[63372]: DEBUG oslo_concurrency.lockutils [req-15d8a8b6-244e-41af-a599-49e4d17afaef req-7d265f98-2acc-493a-9755-ff58e3225c68 service nova] Lock "7a75a51d-b804-453b-ba7b-fda6c4931802-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.931616] env[63372]: DEBUG nova.compute.manager [req-15d8a8b6-244e-41af-a599-49e4d17afaef req-7d265f98-2acc-493a-9755-ff58e3225c68 service nova] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] No waiting events found dispatching network-vif-plugged-6c9c3162-8734-4911-a9d1-9f3f9214ea93 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 737.931616] env[63372]: WARNING nova.compute.manager [req-15d8a8b6-244e-41af-a599-49e4d17afaef req-7d265f98-2acc-493a-9755-ff58e3225c68 service nova] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Received unexpected event network-vif-plugged-6c9c3162-8734-4911-a9d1-9f3f9214ea93 for instance with vm_state building and task_state spawning. [ 737.936849] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023917, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.034450] env[63372]: DEBUG nova.network.neutron [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Successfully updated port: 6c9c3162-8734-4911-a9d1-9f3f9214ea93 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 738.279031] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba88d64-636e-4769-ab7c-2edd40781ba0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.288380] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843eb4e1-4bc8-41cf-a0ee-b2f9a5976e56 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.321367] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a845c52d-fa13-4168-b3a1-52d141ab7276 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.329518] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43cfa367-821c-4eba-a517-ee2d9ea17e2e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.347273] env[63372]: DEBUG nova.compute.provider_tree [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.348625] env[63372]: DEBUG nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 738.435112] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023917, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117052} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.435543] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 738.437064] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff5baf2-ddf6-4cec-a613-8210676b04c5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.467235] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] f43555ef-b517-4b7d-9d2e-4787d40e201b/f43555ef-b517-4b7d-9d2e-4787d40e201b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 738.467414] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-420e4940-8b34-4b03-86d2-0f18f18c543e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.490672] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 738.490672] env[63372]: value = "task-1023918" [ 738.490672] env[63372]: _type = "Task" [ 738.490672] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.500969] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023918, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.538271] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Acquiring lock "refresh_cache-7a75a51d-b804-453b-ba7b-fda6c4931802" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.538544] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Acquired lock "refresh_cache-7a75a51d-b804-453b-ba7b-fda6c4931802" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.539496] env[63372]: DEBUG nova.network.neutron [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 738.577300] env[63372]: DEBUG oslo_concurrency.lockutils [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "63b9095c-fb49-4dc1-a6dc-96529aeaab81" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.577640] env[63372]: DEBUG oslo_concurrency.lockutils [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "63b9095c-fb49-4dc1-a6dc-96529aeaab81" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.577882] env[63372]: DEBUG oslo_concurrency.lockutils [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "63b9095c-fb49-4dc1-a6dc-96529aeaab81-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.578107] env[63372]: DEBUG oslo_concurrency.lockutils [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "63b9095c-fb49-4dc1-a6dc-96529aeaab81-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.578621] env[63372]: DEBUG oslo_concurrency.lockutils [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "63b9095c-fb49-4dc1-a6dc-96529aeaab81-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.583058] env[63372]: INFO nova.compute.manager [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Terminating instance [ 738.584017] env[63372]: DEBUG nova.compute.manager [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 738.584515] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 738.588775] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e55666ca-2c4b-4263-b737-c21dd1c342b0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.597052] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 738.597052] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-844a5a8b-cfcb-41dd-bae1-f8e3fedf7432 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.603614] env[63372]: DEBUG oslo_vmware.api [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 738.603614] env[63372]: value = "task-1023919" [ 738.603614] env[63372]: _type = "Task" [ 738.603614] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.614156] env[63372]: DEBUG oslo_vmware.api [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023919, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.852970] env[63372]: DEBUG nova.scheduler.client.report [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.875462] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.002963] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.108763] env[63372]: DEBUG nova.network.neutron [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.116360] env[63372]: DEBUG oslo_vmware.api [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023919, 'name': PowerOffVM_Task, 'duration_secs': 0.227143} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.116692] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 739.116876] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 739.117900] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0e044be9-54d6-4f7a-8f2a-11d2f230ba00 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.192694] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 739.192986] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 739.193194] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Deleting the datastore file [datastore1] 63b9095c-fb49-4dc1-a6dc-96529aeaab81 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 739.193489] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b77960a0-97b9-43a1-b075-d9448fe7656b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.202631] env[63372]: DEBUG oslo_vmware.api [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 739.202631] env[63372]: value = "task-1023921" [ 739.202631] env[63372]: _type = "Task" [ 739.202631] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.216816] env[63372]: DEBUG oslo_vmware.api [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023921, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.360296] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.583s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.361185] env[63372]: DEBUG nova.compute.manager [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 739.364327] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.190s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.365441] env[63372]: INFO nova.compute.claims [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.438210] env[63372]: DEBUG nova.network.neutron [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Updating instance_info_cache with network_info: [{"id": "6c9c3162-8734-4911-a9d1-9f3f9214ea93", "address": "fa:16:3e:81:26:4e", "network": {"id": "6a147c66-cd17-47c3-993b-bd3aab62811b", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-786554058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9946a38bddee4f7cb2c6551d76f0a1c0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb3425ea-72e7-41e3-92a7-820db9ec4661", "external-id": "nsx-vlan-transportzone-651", "segmentation_id": 651, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9c3162-87", "ovs_interfaceid": "6c9c3162-8734-4911-a9d1-9f3f9214ea93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.503198] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023918, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.715064] env[63372]: DEBUG oslo_vmware.api [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1023921, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.220107} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.715392] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 739.715623] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 739.715845] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 739.716115] env[63372]: INFO nova.compute.manager [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Took 1.13 seconds to destroy the instance on the hypervisor. [ 739.716478] env[63372]: DEBUG oslo.service.loopingcall [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.716744] env[63372]: DEBUG nova.compute.manager [-] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.716911] env[63372]: DEBUG nova.network.neutron [-] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 739.869536] env[63372]: DEBUG nova.compute.utils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 739.871059] env[63372]: DEBUG nova.compute.manager [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 739.871237] env[63372]: DEBUG nova.network.neutron [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 739.941900] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Releasing lock "refresh_cache-7a75a51d-b804-453b-ba7b-fda6c4931802" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.942284] env[63372]: DEBUG nova.compute.manager [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Instance network_info: |[{"id": "6c9c3162-8734-4911-a9d1-9f3f9214ea93", "address": "fa:16:3e:81:26:4e", "network": {"id": "6a147c66-cd17-47c3-993b-bd3aab62811b", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-786554058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9946a38bddee4f7cb2c6551d76f0a1c0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb3425ea-72e7-41e3-92a7-820db9ec4661", "external-id": "nsx-vlan-transportzone-651", "segmentation_id": 651, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9c3162-87", "ovs_interfaceid": "6c9c3162-8734-4911-a9d1-9f3f9214ea93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 739.943859] env[63372]: DEBUG nova.policy [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '053f2ae589f5418785bbb783b635e0a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a14a1aa8a0574b65a312754ae1091f37', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 739.946091] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:26:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eb3425ea-72e7-41e3-92a7-820db9ec4661', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6c9c3162-8734-4911-a9d1-9f3f9214ea93', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 739.953714] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Creating folder: Project (9946a38bddee4f7cb2c6551d76f0a1c0). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 739.954298] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-faa8b1d2-feaf-4a3a-a185-7c4fdb0f2792 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.966256] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Created folder: Project (9946a38bddee4f7cb2c6551d76f0a1c0) in parent group-v227230. [ 739.966467] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Creating folder: Instances. Parent ref: group-v227317. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 739.967903] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8533212f-cf8a-4dd7-9154-c01f8be4bfbd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.976211] env[63372]: DEBUG nova.compute.manager [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Received event network-changed-6c9c3162-8734-4911-a9d1-9f3f9214ea93 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 739.976417] env[63372]: DEBUG nova.compute.manager [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Refreshing instance network info cache due to event network-changed-6c9c3162-8734-4911-a9d1-9f3f9214ea93. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 739.976661] env[63372]: DEBUG oslo_concurrency.lockutils [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] Acquiring lock "refresh_cache-7a75a51d-b804-453b-ba7b-fda6c4931802" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.976814] env[63372]: DEBUG oslo_concurrency.lockutils [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] Acquired lock "refresh_cache-7a75a51d-b804-453b-ba7b-fda6c4931802" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.977042] env[63372]: DEBUG nova.network.neutron [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Refreshing network info cache for port 6c9c3162-8734-4911-a9d1-9f3f9214ea93 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 739.987197] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Created folder: Instances in parent group-v227317. [ 739.988015] env[63372]: DEBUG oslo.service.loopingcall [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.988015] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 739.988015] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c9dd724-e711-4f27-ad52-df180eb75c21 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.012566] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 740.012566] env[63372]: value = "task-1023924" [ 740.012566] env[63372]: _type = "Task" [ 740.012566] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.015722] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023918, 'name': ReconfigVM_Task, 'duration_secs': 1.040899} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.019171] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Reconfigured VM instance instance-00000031 to attach disk [datastore2] f43555ef-b517-4b7d-9d2e-4787d40e201b/f43555ef-b517-4b7d-9d2e-4787d40e201b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 740.019992] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-476d006a-25f7-4a6f-8065-8036d5d4fb84 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.029779] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023924, 'name': CreateVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.031488] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 740.031488] env[63372]: value = "task-1023925" [ 740.031488] env[63372]: _type = "Task" [ 740.031488] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.040277] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023925, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.374405] env[63372]: DEBUG nova.compute.manager [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 740.392285] env[63372]: DEBUG nova.network.neutron [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Successfully created port: 263cc51f-62e9-4a41-9d93-6f8c90cb320d {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 740.528787] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023924, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.546238] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023925, 'name': Rename_Task, 'duration_secs': 0.220852} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.546238] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 740.546238] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8473289-50b5-4c68-8aa3-32dce86cf533 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.554140] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 740.554140] env[63372]: value = "task-1023926" [ 740.554140] env[63372]: _type = "Task" [ 740.554140] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.562141] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023926, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.573193] env[63372]: DEBUG nova.network.neutron [-] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.845227] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4232eefc-eccc-48d3-8a9d-e8c4f1425c9e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.852933] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f1c35b-4777-4494-a6dd-d317841f5bd8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.891015] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2100784-ef4d-40a2-958c-fdc26c63eef6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.898766] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d03311-4392-4cbf-859c-f9ebb8585a64 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.914769] env[63372]: DEBUG nova.compute.provider_tree [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.939386] env[63372]: DEBUG nova.network.neutron [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Updated VIF entry in instance network info cache for port 6c9c3162-8734-4911-a9d1-9f3f9214ea93. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 740.939845] env[63372]: DEBUG nova.network.neutron [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Updating instance_info_cache with network_info: [{"id": "6c9c3162-8734-4911-a9d1-9f3f9214ea93", "address": "fa:16:3e:81:26:4e", "network": {"id": "6a147c66-cd17-47c3-993b-bd3aab62811b", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-786554058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9946a38bddee4f7cb2c6551d76f0a1c0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb3425ea-72e7-41e3-92a7-820db9ec4661", "external-id": "nsx-vlan-transportzone-651", "segmentation_id": 651, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9c3162-87", "ovs_interfaceid": "6c9c3162-8734-4911-a9d1-9f3f9214ea93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.025858] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023924, 'name': CreateVM_Task, 'duration_secs': 0.573672} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.026066] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 741.026818] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.026997] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.027380] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 741.027668] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25c7f7ec-f943-4e33-a5e8-710cab5b403e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.033612] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Waiting for the task: (returnval){ [ 741.033612] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52534c08-e2b7-881a-4252-09bbdb32b6fd" [ 741.033612] env[63372]: _type = "Task" [ 741.033612] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.042026] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52534c08-e2b7-881a-4252-09bbdb32b6fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.064498] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023926, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.076186] env[63372]: INFO nova.compute.manager [-] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Took 1.36 seconds to deallocate network for instance. [ 741.395426] env[63372]: DEBUG nova.compute.manager [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 741.418104] env[63372]: DEBUG nova.scheduler.client.report [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.428082] env[63372]: DEBUG nova.virt.hardware [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 741.428433] env[63372]: DEBUG nova.virt.hardware [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 741.428615] env[63372]: DEBUG nova.virt.hardware [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 741.428812] env[63372]: DEBUG nova.virt.hardware [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 741.428990] env[63372]: DEBUG nova.virt.hardware [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 741.429830] env[63372]: DEBUG nova.virt.hardware [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 741.430125] env[63372]: DEBUG nova.virt.hardware [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 741.430335] env[63372]: DEBUG nova.virt.hardware [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 741.430541] env[63372]: DEBUG nova.virt.hardware [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 741.430813] env[63372]: DEBUG nova.virt.hardware [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 741.431091] env[63372]: DEBUG nova.virt.hardware [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.431910] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1289aeb4-9b66-4b6e-8c0f-4dcc649f57be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.446119] env[63372]: DEBUG oslo_concurrency.lockutils [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] Releasing lock "refresh_cache-7a75a51d-b804-453b-ba7b-fda6c4931802" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.446440] env[63372]: DEBUG nova.compute.manager [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Received event network-changed-8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 741.446645] env[63372]: DEBUG nova.compute.manager [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Refreshing instance network info cache due to event network-changed-8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 741.446877] env[63372]: DEBUG oslo_concurrency.lockutils [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] Acquiring lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.447036] env[63372]: DEBUG oslo_concurrency.lockutils [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] Acquired lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.447231] env[63372]: DEBUG nova.network.neutron [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Refreshing network info cache for port 8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 741.450145] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1182e259-9092-4553-9891-efecd32b3463 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.544754] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52534c08-e2b7-881a-4252-09bbdb32b6fd, 'name': SearchDatastore_Task, 'duration_secs': 0.012994} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.545166] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.545443] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 741.545693] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.545846] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.546059] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 741.546369] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-65ea63f1-69f4-41bf-a8ba-a559ff7196e6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.556358] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 741.556638] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 741.560566] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8751825-a819-48a4-bdf6-2daf4b6befb6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.569557] env[63372]: DEBUG oslo_vmware.api [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023926, 'name': PowerOnVM_Task, 'duration_secs': 0.747163} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.571068] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 741.571208] env[63372]: INFO nova.compute.manager [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Took 10.29 seconds to spawn the instance on the hypervisor. [ 741.571391] env[63372]: DEBUG nova.compute.manager [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 741.571758] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Waiting for the task: (returnval){ [ 741.571758] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525f8d38-c49d-1332-c970-6fc841c3f934" [ 741.571758] env[63372]: _type = "Task" [ 741.571758] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.572590] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ad2e03-64be-4498-a435-3df88746bd7e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.594026] env[63372]: DEBUG oslo_concurrency.lockutils [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.594196] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525f8d38-c49d-1332-c970-6fc841c3f934, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.856813] env[63372]: DEBUG nova.compute.manager [req-022e456d-fdd2-439a-8ebf-e69ebd241b93 req-83784927-bc2e-4f3f-a641-6607921bf744 service nova] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Received event network-vif-plugged-263cc51f-62e9-4a41-9d93-6f8c90cb320d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 741.857068] env[63372]: DEBUG oslo_concurrency.lockutils [req-022e456d-fdd2-439a-8ebf-e69ebd241b93 req-83784927-bc2e-4f3f-a641-6607921bf744 service nova] Acquiring lock "cd0c01ac-602b-44a3-8099-84b8a50b2449-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.857287] env[63372]: DEBUG oslo_concurrency.lockutils [req-022e456d-fdd2-439a-8ebf-e69ebd241b93 req-83784927-bc2e-4f3f-a641-6607921bf744 service nova] Lock "cd0c01ac-602b-44a3-8099-84b8a50b2449-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.857458] env[63372]: DEBUG oslo_concurrency.lockutils [req-022e456d-fdd2-439a-8ebf-e69ebd241b93 req-83784927-bc2e-4f3f-a641-6607921bf744 service nova] Lock "cd0c01ac-602b-44a3-8099-84b8a50b2449-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.857629] env[63372]: DEBUG nova.compute.manager [req-022e456d-fdd2-439a-8ebf-e69ebd241b93 req-83784927-bc2e-4f3f-a641-6607921bf744 service nova] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] No waiting events found dispatching network-vif-plugged-263cc51f-62e9-4a41-9d93-6f8c90cb320d {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 741.857856] env[63372]: WARNING nova.compute.manager [req-022e456d-fdd2-439a-8ebf-e69ebd241b93 req-83784927-bc2e-4f3f-a641-6607921bf744 service nova] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Received unexpected event network-vif-plugged-263cc51f-62e9-4a41-9d93-6f8c90cb320d for instance with vm_state building and task_state spawning. [ 741.942024] env[63372]: DEBUG nova.network.neutron [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Successfully updated port: 263cc51f-62e9-4a41-9d93-6f8c90cb320d {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 741.942024] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.577s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.942024] env[63372]: DEBUG nova.compute.manager [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 741.945237] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.049s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.949579] env[63372]: DEBUG nova.objects.instance [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lazy-loading 'resources' on Instance uuid e5522005-8c54-43e1-ae23-5e5ff1ef0ee9 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 742.040369] env[63372]: DEBUG nova.compute.manager [req-a8fa0c7d-f994-45ff-9750-2d7b3c08fd94 req-583d54e6-b764-49a9-ba86-f7b24f15fe3f service nova] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Received event network-vif-deleted-022ca5fc-1544-4248-bb33-e377ab46b598 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.105925] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525f8d38-c49d-1332-c970-6fc841c3f934, 'name': SearchDatastore_Task, 'duration_secs': 0.015134} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.106489] env[63372]: INFO nova.compute.manager [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Took 39.83 seconds to build instance. [ 742.108574] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8f9e93e-48cf-4028-92cc-3cfc79ab7c50 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.117783] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Waiting for the task: (returnval){ [ 742.117783] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]527b0819-fc29-92a4-bc0a-2f51cfa2ef15" [ 742.117783] env[63372]: _type = "Task" [ 742.117783] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.128555] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527b0819-fc29-92a4-bc0a-2f51cfa2ef15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.289291] env[63372]: DEBUG nova.network.neutron [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updated VIF entry in instance network info cache for port 8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 742.289689] env[63372]: DEBUG nova.network.neutron [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance_info_cache with network_info: [{"id": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "address": "fa:16:3e:3f:99:77", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c4c74ba-ad", "ovs_interfaceid": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.441188] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "refresh_cache-cd0c01ac-602b-44a3-8099-84b8a50b2449" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.441492] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquired lock "refresh_cache-cd0c01ac-602b-44a3-8099-84b8a50b2449" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.441492] env[63372]: DEBUG nova.network.neutron [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 742.448100] env[63372]: DEBUG nova.compute.utils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 742.448100] env[63372]: DEBUG nova.compute.manager [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 742.448100] env[63372]: DEBUG nova.network.neutron [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 742.492419] env[63372]: DEBUG nova.policy [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f8e364e3aed40248cc0d11741f76154', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b4548b52cd704cc0b054e2f7d1562b9a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 742.610864] env[63372]: DEBUG oslo_concurrency.lockutils [None req-82d1e103-6efe-41ce-89ab-db781fefaba7 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Lock "f43555ef-b517-4b7d-9d2e-4787d40e201b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.738s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.631392] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527b0819-fc29-92a4-bc0a-2f51cfa2ef15, 'name': SearchDatastore_Task, 'duration_secs': 0.013158} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.631392] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.631655] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 7a75a51d-b804-453b-ba7b-fda6c4931802/7a75a51d-b804-453b-ba7b-fda6c4931802.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 742.638179] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e4e32118-1057-49d1-bcda-b9f9d62e108d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.646299] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Waiting for the task: (returnval){ [ 742.646299] env[63372]: value = "task-1023927" [ 742.646299] env[63372]: _type = "Task" [ 742.646299] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.657446] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1023927, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.792492] env[63372]: DEBUG oslo_concurrency.lockutils [req-b657dd5c-71b4-4ea5-8dbd-8413bf0f734c req-7d490471-98fc-4ef9-a6b5-400eac334774 service nova] Releasing lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.810360] env[63372]: DEBUG nova.network.neutron [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Successfully created port: 18da52d3-9658-4af9-bd78-16e5ab3be9d9 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.921021] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f565bfb4-862b-44bb-8510-5c33e499f58a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.928115] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf8e25e-f59d-4708-89bf-a7af9000ae75 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.974392] env[63372]: DEBUG nova.compute.manager [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 742.981019] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7db0f13-68c6-47eb-933b-67fc1a7c2340 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.993868] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df98e10-254c-420e-9a14-59a455f58c91 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.014948] env[63372]: DEBUG nova.compute.provider_tree [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.075840] env[63372]: DEBUG nova.network.neutron [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.114011] env[63372]: DEBUG nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 743.162054] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1023927, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.326110] env[63372]: DEBUG nova.network.neutron [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Updating instance_info_cache with network_info: [{"id": "263cc51f-62e9-4a41-9d93-6f8c90cb320d", "address": "fa:16:3e:80:76:5b", "network": {"id": "2f241a45-cc18-46f6-b0a3-1aad8d2713fd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1953733697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a14a1aa8a0574b65a312754ae1091f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a88b707-352e-4be7-b1d6-ad6074b40ed9", "external-id": "nsx-vlan-transportzone-789", "segmentation_id": 789, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap263cc51f-62", "ovs_interfaceid": "263cc51f-62e9-4a41-9d93-6f8c90cb320d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.408829] env[63372]: DEBUG nova.compute.manager [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 743.409811] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c88c36-e11a-4d47-aca5-ba8e4e67b4a9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.518457] env[63372]: DEBUG nova.scheduler.client.report [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.637886] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.658192] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1023927, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.621104} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.659047] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 7a75a51d-b804-453b-ba7b-fda6c4931802/7a75a51d-b804-453b-ba7b-fda6c4931802.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 743.659047] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 743.659305] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a3f46dbc-7d41-4f75-bcb7-e4c52ccd12f7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.666455] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Waiting for the task: (returnval){ [ 743.666455] env[63372]: value = "task-1023928" [ 743.666455] env[63372]: _type = "Task" [ 743.666455] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.674854] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1023928, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.829747] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Releasing lock "refresh_cache-cd0c01ac-602b-44a3-8099-84b8a50b2449" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.829747] env[63372]: DEBUG nova.compute.manager [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Instance network_info: |[{"id": "263cc51f-62e9-4a41-9d93-6f8c90cb320d", "address": "fa:16:3e:80:76:5b", "network": {"id": "2f241a45-cc18-46f6-b0a3-1aad8d2713fd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1953733697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a14a1aa8a0574b65a312754ae1091f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a88b707-352e-4be7-b1d6-ad6074b40ed9", "external-id": "nsx-vlan-transportzone-789", "segmentation_id": 789, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap263cc51f-62", "ovs_interfaceid": "263cc51f-62e9-4a41-9d93-6f8c90cb320d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 743.830266] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:76:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0a88b707-352e-4be7-b1d6-ad6074b40ed9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '263cc51f-62e9-4a41-9d93-6f8c90cb320d', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 743.840549] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Creating folder: Project (a14a1aa8a0574b65a312754ae1091f37). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 743.841349] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aa61fa26-ad40-40dc-bb7a-be708c382ec4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.852176] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Created folder: Project (a14a1aa8a0574b65a312754ae1091f37) in parent group-v227230. [ 743.852420] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Creating folder: Instances. Parent ref: group-v227320. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 743.852641] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5960af54-e68e-4a57-9a42-8c2ec68525af {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.862474] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Created folder: Instances in parent group-v227320. [ 743.862474] env[63372]: DEBUG oslo.service.loopingcall [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.862474] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 743.862474] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-79b78bb4-04be-410f-8b93-fa33027d184c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.881683] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 743.881683] env[63372]: value = "task-1023931" [ 743.881683] env[63372]: _type = "Task" [ 743.881683] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.889580] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023931, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.914228] env[63372]: DEBUG nova.compute.manager [req-35262a2b-cc24-47ba-b841-58b561ade152 req-21d648c9-2ab9-48ca-9fd0-18ed5b4ca9e5 service nova] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Received event network-changed-263cc51f-62e9-4a41-9d93-6f8c90cb320d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 743.914477] env[63372]: DEBUG nova.compute.manager [req-35262a2b-cc24-47ba-b841-58b561ade152 req-21d648c9-2ab9-48ca-9fd0-18ed5b4ca9e5 service nova] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Refreshing instance network info cache due to event network-changed-263cc51f-62e9-4a41-9d93-6f8c90cb320d. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 743.914722] env[63372]: DEBUG oslo_concurrency.lockutils [req-35262a2b-cc24-47ba-b841-58b561ade152 req-21d648c9-2ab9-48ca-9fd0-18ed5b4ca9e5 service nova] Acquiring lock "refresh_cache-cd0c01ac-602b-44a3-8099-84b8a50b2449" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.914897] env[63372]: DEBUG oslo_concurrency.lockutils [req-35262a2b-cc24-47ba-b841-58b561ade152 req-21d648c9-2ab9-48ca-9fd0-18ed5b4ca9e5 service nova] Acquired lock "refresh_cache-cd0c01ac-602b-44a3-8099-84b8a50b2449" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.915081] env[63372]: DEBUG nova.network.neutron [req-35262a2b-cc24-47ba-b841-58b561ade152 req-21d648c9-2ab9-48ca-9fd0-18ed5b4ca9e5 service nova] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Refreshing network info cache for port 263cc51f-62e9-4a41-9d93-6f8c90cb320d {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 743.920907] env[63372]: INFO nova.compute.manager [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] instance snapshotting [ 743.924216] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f1eabc9-470f-41f1-9d05-4220fd15ff2c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.950077] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8686d2-e054-42c5-9a3e-32ae4104c0f3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.986826] env[63372]: DEBUG nova.compute.manager [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 744.009857] env[63372]: DEBUG nova.virt.hardware [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 744.010129] env[63372]: DEBUG nova.virt.hardware [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 744.010293] env[63372]: DEBUG nova.virt.hardware [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 744.010491] env[63372]: DEBUG nova.virt.hardware [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 744.010719] env[63372]: DEBUG nova.virt.hardware [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 744.010943] env[63372]: DEBUG nova.virt.hardware [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 744.011258] env[63372]: DEBUG nova.virt.hardware [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 744.011472] env[63372]: DEBUG nova.virt.hardware [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 744.011681] env[63372]: DEBUG nova.virt.hardware [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 744.011915] env[63372]: DEBUG nova.virt.hardware [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 744.012174] env[63372]: DEBUG nova.virt.hardware [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 744.013297] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21935f90-8384-43fb-9a05-20a622edb391 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.022039] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea2e2b8-3e05-4176-bec5-2ac0b1e14d2f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.026778] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.082s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.028981] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.487s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.032091] env[63372]: INFO nova.compute.claims [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.051530] env[63372]: INFO nova.scheduler.client.report [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Deleted allocations for instance e5522005-8c54-43e1-ae23-5e5ff1ef0ee9 [ 744.178115] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1023928, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.119132} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.178115] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 744.178989] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6d6ebe-11d4-480b-a7a4-d907b8a8711f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.207850] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 7a75a51d-b804-453b-ba7b-fda6c4931802/7a75a51d-b804-453b-ba7b-fda6c4931802.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 744.207971] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aeba4cb3-68da-40de-9966-d22cb8bade8a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.233010] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Waiting for the task: (returnval){ [ 744.233010] env[63372]: value = "task-1023932" [ 744.233010] env[63372]: _type = "Task" [ 744.233010] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.243551] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1023932, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.391972] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023931, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.460932] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 744.461332] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-865b67ce-9ee3-46e0-a098-b480f4a90671 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.470057] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 744.470057] env[63372]: value = "task-1023933" [ 744.470057] env[63372]: _type = "Task" [ 744.470057] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.479395] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023933, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.534074] env[63372]: DEBUG nova.network.neutron [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Successfully updated port: 18da52d3-9658-4af9-bd78-16e5ab3be9d9 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 744.559901] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd24f27b-835d-47ee-bafd-37578da63123 tempest-SecurityGroupsTestJSON-436381323 tempest-SecurityGroupsTestJSON-436381323-project-member] Lock "e5522005-8c54-43e1-ae23-5e5ff1ef0ee9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.704s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.673048] env[63372]: DEBUG nova.network.neutron [req-35262a2b-cc24-47ba-b841-58b561ade152 req-21d648c9-2ab9-48ca-9fd0-18ed5b4ca9e5 service nova] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Updated VIF entry in instance network info cache for port 263cc51f-62e9-4a41-9d93-6f8c90cb320d. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 744.673468] env[63372]: DEBUG nova.network.neutron [req-35262a2b-cc24-47ba-b841-58b561ade152 req-21d648c9-2ab9-48ca-9fd0-18ed5b4ca9e5 service nova] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Updating instance_info_cache with network_info: [{"id": "263cc51f-62e9-4a41-9d93-6f8c90cb320d", "address": "fa:16:3e:80:76:5b", "network": {"id": "2f241a45-cc18-46f6-b0a3-1aad8d2713fd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1953733697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a14a1aa8a0574b65a312754ae1091f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a88b707-352e-4be7-b1d6-ad6074b40ed9", "external-id": "nsx-vlan-transportzone-789", "segmentation_id": 789, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap263cc51f-62", "ovs_interfaceid": "263cc51f-62e9-4a41-9d93-6f8c90cb320d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.749138] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1023932, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.892297] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023931, 'name': CreateVM_Task, 'duration_secs': 0.657806} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.892521] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 744.893255] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.893433] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.893803] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 744.894124] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea972047-4a8f-4281-a9ca-54d848dafeba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.899104] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 744.899104] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52daeefb-bbee-ffbd-1fb3-1c0b6805c31d" [ 744.899104] env[63372]: _type = "Task" [ 744.899104] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.912443] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52daeefb-bbee-ffbd-1fb3-1c0b6805c31d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.982910] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023933, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.045711] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "refresh_cache-ae6f3f42-7213-4ab1-b74c-1a557df6748b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.045711] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired lock "refresh_cache-ae6f3f42-7213-4ab1-b74c-1a557df6748b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.045711] env[63372]: DEBUG nova.network.neutron [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 745.176568] env[63372]: DEBUG oslo_concurrency.lockutils [req-35262a2b-cc24-47ba-b841-58b561ade152 req-21d648c9-2ab9-48ca-9fd0-18ed5b4ca9e5 service nova] Releasing lock "refresh_cache-cd0c01ac-602b-44a3-8099-84b8a50b2449" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.246700] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1023932, 'name': ReconfigVM_Task, 'duration_secs': 0.682157} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.247061] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 7a75a51d-b804-453b-ba7b-fda6c4931802/7a75a51d-b804-453b-ba7b-fda6c4931802.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 745.247732] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c750dfb6-03a2-4098-9d78-5fa285121609 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.255054] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Waiting for the task: (returnval){ [ 745.255054] env[63372]: value = "task-1023934" [ 745.255054] env[63372]: _type = "Task" [ 745.255054] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.264062] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1023934, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.413352] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52daeefb-bbee-ffbd-1fb3-1c0b6805c31d, 'name': SearchDatastore_Task, 'duration_secs': 0.017874} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.413601] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.413850] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 745.414137] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.414381] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.414478] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 745.414739] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9bb5fb70-c2ff-468a-a6f3-9ddbec7d7408 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.427090] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 745.427324] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 745.428184] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e9fa7fc-f115-4664-9d43-29a26ffee424 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.436653] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 745.436653] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52154c27-ea71-ae7e-12ff-e58d3927950e" [ 745.436653] env[63372]: _type = "Task" [ 745.436653] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.445319] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52154c27-ea71-ae7e-12ff-e58d3927950e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.477017] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96986bd2-0122-4a63-bc5b-650f2bd7a2ac {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.485056] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023933, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.487961] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3bc9e3-c38b-40aa-8f7d-0b8ff464aada {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.519024] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43633e3a-865d-4fe0-ae37-1e442838f553 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.527125] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6907d25-9a27-4b03-8d7e-07a2a81a4111 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.543801] env[63372]: DEBUG nova.compute.provider_tree [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.608343] env[63372]: DEBUG nova.network.neutron [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 745.768678] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1023934, 'name': Rename_Task, 'duration_secs': 0.200791} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.768678] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 745.768678] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-930924b8-d58f-4319-936d-0287d8be0020 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.776297] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Waiting for the task: (returnval){ [ 745.776297] env[63372]: value = "task-1023935" [ 745.776297] env[63372]: _type = "Task" [ 745.776297] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.784218] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1023935, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.919489] env[63372]: DEBUG nova.network.neutron [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Updating instance_info_cache with network_info: [{"id": "18da52d3-9658-4af9-bd78-16e5ab3be9d9", "address": "fa:16:3e:c7:40:3c", "network": {"id": "7157ae73-79e9-454e-8673-7841f4cfe1f5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-786035586-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4548b52cd704cc0b054e2f7d1562b9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18da52d3-96", "ovs_interfaceid": "18da52d3-9658-4af9-bd78-16e5ab3be9d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.948463] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52154c27-ea71-ae7e-12ff-e58d3927950e, 'name': SearchDatastore_Task, 'duration_secs': 0.010412} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.949344] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da6ffec2-2ef6-44ee-b27d-b2a7352a08f4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.955027] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 745.955027] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525da7c8-2dd6-4877-1660-70df49b5b7e7" [ 745.955027] env[63372]: _type = "Task" [ 745.955027] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.963799] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525da7c8-2dd6-4877-1660-70df49b5b7e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.981494] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023933, 'name': CreateSnapshot_Task, 'duration_secs': 1.102592} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.981818] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 745.982846] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f5f038b-b02d-49fa-9eaa-77770ee4555f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.023675] env[63372]: DEBUG nova.compute.manager [req-c70aefd3-d12c-4d17-aab1-08a10b77585d req-fe88667b-b77d-43bc-8508-d3cc5a43bb21 service nova] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Received event network-vif-plugged-18da52d3-9658-4af9-bd78-16e5ab3be9d9 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.023730] env[63372]: DEBUG oslo_concurrency.lockutils [req-c70aefd3-d12c-4d17-aab1-08a10b77585d req-fe88667b-b77d-43bc-8508-d3cc5a43bb21 service nova] Acquiring lock "ae6f3f42-7213-4ab1-b74c-1a557df6748b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.024048] env[63372]: DEBUG oslo_concurrency.lockutils [req-c70aefd3-d12c-4d17-aab1-08a10b77585d req-fe88667b-b77d-43bc-8508-d3cc5a43bb21 service nova] Lock "ae6f3f42-7213-4ab1-b74c-1a557df6748b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.024230] env[63372]: DEBUG oslo_concurrency.lockutils [req-c70aefd3-d12c-4d17-aab1-08a10b77585d req-fe88667b-b77d-43bc-8508-d3cc5a43bb21 service nova] Lock "ae6f3f42-7213-4ab1-b74c-1a557df6748b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.024452] env[63372]: DEBUG nova.compute.manager [req-c70aefd3-d12c-4d17-aab1-08a10b77585d req-fe88667b-b77d-43bc-8508-d3cc5a43bb21 service nova] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] No waiting events found dispatching network-vif-plugged-18da52d3-9658-4af9-bd78-16e5ab3be9d9 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 746.024672] env[63372]: WARNING nova.compute.manager [req-c70aefd3-d12c-4d17-aab1-08a10b77585d req-fe88667b-b77d-43bc-8508-d3cc5a43bb21 service nova] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Received unexpected event network-vif-plugged-18da52d3-9658-4af9-bd78-16e5ab3be9d9 for instance with vm_state building and task_state spawning. [ 746.024857] env[63372]: DEBUG nova.compute.manager [req-c70aefd3-d12c-4d17-aab1-08a10b77585d req-fe88667b-b77d-43bc-8508-d3cc5a43bb21 service nova] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Received event network-changed-18da52d3-9658-4af9-bd78-16e5ab3be9d9 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.025034] env[63372]: DEBUG nova.compute.manager [req-c70aefd3-d12c-4d17-aab1-08a10b77585d req-fe88667b-b77d-43bc-8508-d3cc5a43bb21 service nova] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Refreshing instance network info cache due to event network-changed-18da52d3-9658-4af9-bd78-16e5ab3be9d9. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 746.025236] env[63372]: DEBUG oslo_concurrency.lockutils [req-c70aefd3-d12c-4d17-aab1-08a10b77585d req-fe88667b-b77d-43bc-8508-d3cc5a43bb21 service nova] Acquiring lock "refresh_cache-ae6f3f42-7213-4ab1-b74c-1a557df6748b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.047198] env[63372]: DEBUG nova.scheduler.client.report [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.287788] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1023935, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.293935] env[63372]: DEBUG oslo_vmware.rw_handles [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5206ac3d-20d3-e247-b36b-ab82c883fb3a/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 746.295303] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03584f28-3298-46ba-b7ba-f19e2d18f36a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.301959] env[63372]: DEBUG oslo_vmware.rw_handles [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5206ac3d-20d3-e247-b36b-ab82c883fb3a/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 746.301959] env[63372]: ERROR oslo_vmware.rw_handles [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5206ac3d-20d3-e247-b36b-ab82c883fb3a/disk-0.vmdk due to incomplete transfer. [ 746.301959] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5cd515cb-f04c-4910-be8c-eb8dd3f79eae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.310217] env[63372]: DEBUG oslo_vmware.rw_handles [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5206ac3d-20d3-e247-b36b-ab82c883fb3a/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 746.310434] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Uploaded image 184e52b2-673c-4afd-b14d-dcd92d374426 to the Glance image server {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 746.313369] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 746.313639] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8540b319-3eb9-41e8-a86f-0378771643b2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.320156] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 746.320156] env[63372]: value = "task-1023936" [ 746.320156] env[63372]: _type = "Task" [ 746.320156] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.328426] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023936, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.422383] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Releasing lock "refresh_cache-ae6f3f42-7213-4ab1-b74c-1a557df6748b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.422846] env[63372]: DEBUG nova.compute.manager [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Instance network_info: |[{"id": "18da52d3-9658-4af9-bd78-16e5ab3be9d9", "address": "fa:16:3e:c7:40:3c", "network": {"id": "7157ae73-79e9-454e-8673-7841f4cfe1f5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-786035586-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4548b52cd704cc0b054e2f7d1562b9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18da52d3-96", "ovs_interfaceid": "18da52d3-9658-4af9-bd78-16e5ab3be9d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 746.424663] env[63372]: DEBUG oslo_concurrency.lockutils [req-c70aefd3-d12c-4d17-aab1-08a10b77585d req-fe88667b-b77d-43bc-8508-d3cc5a43bb21 service nova] Acquired lock "refresh_cache-ae6f3f42-7213-4ab1-b74c-1a557df6748b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.424663] env[63372]: DEBUG nova.network.neutron [req-c70aefd3-d12c-4d17-aab1-08a10b77585d req-fe88667b-b77d-43bc-8508-d3cc5a43bb21 service nova] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Refreshing network info cache for port 18da52d3-9658-4af9-bd78-16e5ab3be9d9 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 746.425041] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c7:40:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '18da52d3-9658-4af9-bd78-16e5ab3be9d9', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 746.433192] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Creating folder: Project (b4548b52cd704cc0b054e2f7d1562b9a). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 746.433815] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6ab03a5d-190b-4959-b26d-5e0de2acdb75 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.445474] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Created folder: Project (b4548b52cd704cc0b054e2f7d1562b9a) in parent group-v227230. [ 746.446207] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Creating folder: Instances. Parent ref: group-v227324. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 746.446207] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c0d5834-bc55-4852-b7d0-5d1346ee267f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.460413] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Created folder: Instances in parent group-v227324. [ 746.460650] env[63372]: DEBUG oslo.service.loopingcall [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 746.461234] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 746.461445] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fb9d6d06-7f07-4b08-b8af-09b082403710 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.481696] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525da7c8-2dd6-4877-1660-70df49b5b7e7, 'name': SearchDatastore_Task, 'duration_secs': 0.013993} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.481696] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.481997] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] cd0c01ac-602b-44a3-8099-84b8a50b2449/cd0c01ac-602b-44a3-8099-84b8a50b2449.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 746.482159] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba6260e6-2060-4e3f-a4a1-f5176b050855 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.485427] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 746.485427] env[63372]: value = "task-1023939" [ 746.485427] env[63372]: _type = "Task" [ 746.485427] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.490593] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 746.490593] env[63372]: value = "task-1023940" [ 746.490593] env[63372]: _type = "Task" [ 746.490593] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.493750] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023939, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.503844] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 746.504151] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0e8980f9-8a3b-4b46-ada7-fc84eae2d1a2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.512201] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023940, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.513834] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 746.513834] env[63372]: value = "task-1023941" [ 746.513834] env[63372]: _type = "Task" [ 746.513834] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.521861] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023941, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.552967] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.524s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.553690] env[63372]: DEBUG nova.compute.manager [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 746.556963] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.467s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.559106] env[63372]: INFO nova.compute.claims [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.789717] env[63372]: DEBUG oslo_vmware.api [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1023935, 'name': PowerOnVM_Task, 'duration_secs': 0.685903} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.790078] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 746.790299] env[63372]: INFO nova.compute.manager [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Took 10.59 seconds to spawn the instance on the hypervisor. [ 746.790488] env[63372]: DEBUG nova.compute.manager [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 746.791368] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8e21fb-8ad4-438f-8636-34446d714dec {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.832196] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023936, 'name': Destroy_Task, 'duration_secs': 0.351717} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.832502] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Destroyed the VM [ 746.832745] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 746.833027] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-37cacb82-1bf1-4809-9ac2-aff1678d3724 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.842313] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 746.842313] env[63372]: value = "task-1023942" [ 746.842313] env[63372]: _type = "Task" [ 746.842313] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.852609] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023942, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.996476] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023939, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.005465] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023940, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.023790] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023941, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.066147] env[63372]: DEBUG nova.compute.utils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 747.070666] env[63372]: DEBUG nova.compute.manager [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 747.070666] env[63372]: DEBUG nova.network.neutron [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 747.168228] env[63372]: DEBUG nova.policy [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5039f0556904497c8cb69ab549d9fb32', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d4892a06e8b45e1873801404caf0739', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 747.256840] env[63372]: DEBUG nova.network.neutron [req-c70aefd3-d12c-4d17-aab1-08a10b77585d req-fe88667b-b77d-43bc-8508-d3cc5a43bb21 service nova] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Updated VIF entry in instance network info cache for port 18da52d3-9658-4af9-bd78-16e5ab3be9d9. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 747.257363] env[63372]: DEBUG nova.network.neutron [req-c70aefd3-d12c-4d17-aab1-08a10b77585d req-fe88667b-b77d-43bc-8508-d3cc5a43bb21 service nova] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Updating instance_info_cache with network_info: [{"id": "18da52d3-9658-4af9-bd78-16e5ab3be9d9", "address": "fa:16:3e:c7:40:3c", "network": {"id": "7157ae73-79e9-454e-8673-7841f4cfe1f5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-786035586-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4548b52cd704cc0b054e2f7d1562b9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18da52d3-96", "ovs_interfaceid": "18da52d3-9658-4af9-bd78-16e5ab3be9d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.314244] env[63372]: INFO nova.compute.manager [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Took 40.29 seconds to build instance. [ 747.351637] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023942, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.497261] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023939, 'name': CreateVM_Task, 'duration_secs': 0.555896} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.500350] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 747.501133] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.501304] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.501624] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 747.502262] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff99b5b2-af73-4ec5-8fc2-02826ceea886 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.507348] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023940, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.62454} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.507958] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] cd0c01ac-602b-44a3-8099-84b8a50b2449/cd0c01ac-602b-44a3-8099-84b8a50b2449.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 747.508209] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 747.508477] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8bce739-94c5-4252-b966-29768e872b88 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.511307] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 747.511307] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525da4f8-1f04-04fd-86e6-6513255ac6a6" [ 747.511307] env[63372]: _type = "Task" [ 747.511307] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.516219] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 747.516219] env[63372]: value = "task-1023943" [ 747.516219] env[63372]: _type = "Task" [ 747.516219] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.523286] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525da4f8-1f04-04fd-86e6-6513255ac6a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.532450] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023941, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.536062] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023943, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.572659] env[63372]: DEBUG nova.compute.manager [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 747.649613] env[63372]: DEBUG nova.network.neutron [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Successfully created port: efd54f40-2ca1-4d06-a8c7-428f7508abe9 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 747.762184] env[63372]: DEBUG oslo_concurrency.lockutils [req-c70aefd3-d12c-4d17-aab1-08a10b77585d req-fe88667b-b77d-43bc-8508-d3cc5a43bb21 service nova] Releasing lock "refresh_cache-ae6f3f42-7213-4ab1-b74c-1a557df6748b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.818173] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3612e44f-192c-4932-96aa-49761d4799cc tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Lock "7a75a51d-b804-453b-ba7b-fda6c4931802" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.302s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.852807] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023942, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.039966] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023941, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.040107] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525da4f8-1f04-04fd-86e6-6513255ac6a6, 'name': SearchDatastore_Task, 'duration_secs': 0.016685} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.041894] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.042832] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 748.042832] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.042832] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.042832] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 748.043169] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9af2969c-c8a4-41ec-8c57-3d072b5356bd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.047738] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023943, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.212321} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.048814] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 748.049152] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf547b5d-f89b-4035-adcb-77467aa9711a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.076864] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] cd0c01ac-602b-44a3-8099-84b8a50b2449/cd0c01ac-602b-44a3-8099-84b8a50b2449.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 748.081414] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab159418-af73-4d37-9b29-84b59f843213 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.101637] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 748.101939] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 748.104180] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-432bfb2e-9212-4670-80ca-32e76409d57a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.111102] env[63372]: DEBUG nova.compute.manager [req-2132607c-2657-481d-8ebd-597b8734ee90 req-1573c9a9-e301-458e-86af-492035d4314f service nova] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Received event network-changed-6c9c3162-8734-4911-a9d1-9f3f9214ea93 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.111309] env[63372]: DEBUG nova.compute.manager [req-2132607c-2657-481d-8ebd-597b8734ee90 req-1573c9a9-e301-458e-86af-492035d4314f service nova] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Refreshing instance network info cache due to event network-changed-6c9c3162-8734-4911-a9d1-9f3f9214ea93. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 748.111549] env[63372]: DEBUG oslo_concurrency.lockutils [req-2132607c-2657-481d-8ebd-597b8734ee90 req-1573c9a9-e301-458e-86af-492035d4314f service nova] Acquiring lock "refresh_cache-7a75a51d-b804-453b-ba7b-fda6c4931802" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.111715] env[63372]: DEBUG oslo_concurrency.lockutils [req-2132607c-2657-481d-8ebd-597b8734ee90 req-1573c9a9-e301-458e-86af-492035d4314f service nova] Acquired lock "refresh_cache-7a75a51d-b804-453b-ba7b-fda6c4931802" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.111894] env[63372]: DEBUG nova.network.neutron [req-2132607c-2657-481d-8ebd-597b8734ee90 req-1573c9a9-e301-458e-86af-492035d4314f service nova] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Refreshing network info cache for port 6c9c3162-8734-4911-a9d1-9f3f9214ea93 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 748.117232] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 748.117232] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52457b3f-2996-4c88-aff4-b76287f4eab2" [ 748.117232] env[63372]: _type = "Task" [ 748.117232] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.124867] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 748.124867] env[63372]: value = "task-1023944" [ 748.124867] env[63372]: _type = "Task" [ 748.124867] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.136021] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52457b3f-2996-4c88-aff4-b76287f4eab2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.138330] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023944, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.141911] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cface6e-106b-457a-8dbe-c1b9080b44fd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.149586] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2920fc2b-ef46-4185-ba71-a638c1af4186 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.182883] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf6eeb9-47f1-46aa-b462-7032d4edbe2e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.191867] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c75dbf-d078-411e-aab7-cdc5cd8c7d9f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.206470] env[63372]: DEBUG nova.compute.provider_tree [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.321056] env[63372]: DEBUG nova.compute.manager [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 748.356651] env[63372]: DEBUG oslo_vmware.api [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023942, 'name': RemoveSnapshot_Task, 'duration_secs': 1.50439} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.358428] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 748.358428] env[63372]: INFO nova.compute.manager [None req-17ca7d32-b5c5-471d-905b-1c1a34283a0c tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Took 16.64 seconds to snapshot the instance on the hypervisor. [ 748.530268] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023941, 'name': CloneVM_Task} progress is 95%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.614795] env[63372]: DEBUG nova.compute.manager [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 748.628805] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52457b3f-2996-4c88-aff4-b76287f4eab2, 'name': SearchDatastore_Task, 'duration_secs': 0.015045} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.634271] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aad785b4-5846-48a4-a712-1c9378d738e3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.642158] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023944, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.644440] env[63372]: DEBUG nova.virt.hardware [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 748.644657] env[63372]: DEBUG nova.virt.hardware [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 748.644816] env[63372]: DEBUG nova.virt.hardware [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 748.644999] env[63372]: DEBUG nova.virt.hardware [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 748.645164] env[63372]: DEBUG nova.virt.hardware [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 748.645313] env[63372]: DEBUG nova.virt.hardware [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 748.645764] env[63372]: DEBUG nova.virt.hardware [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 748.645764] env[63372]: DEBUG nova.virt.hardware [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 748.645844] env[63372]: DEBUG nova.virt.hardware [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 748.645984] env[63372]: DEBUG nova.virt.hardware [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 748.646279] env[63372]: DEBUG nova.virt.hardware [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 748.648027] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbdcecd5-6fba-4528-9602-fb0e5ca50029 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.650606] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 748.650606] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523f8939-f685-7962-2a05-65124c4fe9e5" [ 748.650606] env[63372]: _type = "Task" [ 748.650606] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.657447] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff7ee62-65a0-4b37-9ef4-894b5fe1f822 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.664833] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523f8939-f685-7962-2a05-65124c4fe9e5, 'name': SearchDatastore_Task, 'duration_secs': 0.015701} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.665439] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.665707] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] ae6f3f42-7213-4ab1-b74c-1a557df6748b/ae6f3f42-7213-4ab1-b74c-1a557df6748b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 748.665953] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-758912d3-889f-49b9-87a8-166b7b3057d5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.683289] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 748.683289] env[63372]: value = "task-1023945" [ 748.683289] env[63372]: _type = "Task" [ 748.683289] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.690679] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023945, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.710472] env[63372]: DEBUG nova.scheduler.client.report [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.847785] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.036724] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023941, 'name': CloneVM_Task, 'duration_secs': 2.146333} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.037509] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Created linked-clone VM from snapshot [ 749.042640] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575522d2-405b-4341-b02d-0a43e0236934 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.053096] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Uploading image 7dfc6004-7365-4eea-b15a-80afc3639b32 {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 749.081939] env[63372]: DEBUG oslo_vmware.rw_handles [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 749.081939] env[63372]: value = "vm-227327" [ 749.081939] env[63372]: _type = "VirtualMachine" [ 749.081939] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 749.082285] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e0603ad7-978f-4189-b78f-117edb8f71ab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.091431] env[63372]: DEBUG oslo_vmware.rw_handles [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Lease: (returnval){ [ 749.091431] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5276d793-eebe-af70-33f4-9e0adb7191f6" [ 749.091431] env[63372]: _type = "HttpNfcLease" [ 749.091431] env[63372]: } obtained for exporting VM: (result){ [ 749.091431] env[63372]: value = "vm-227327" [ 749.091431] env[63372]: _type = "VirtualMachine" [ 749.091431] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 749.091804] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the lease: (returnval){ [ 749.091804] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5276d793-eebe-af70-33f4-9e0adb7191f6" [ 749.091804] env[63372]: _type = "HttpNfcLease" [ 749.091804] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 749.098112] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 749.098112] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5276d793-eebe-af70-33f4-9e0adb7191f6" [ 749.098112] env[63372]: _type = "HttpNfcLease" [ 749.098112] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 749.145876] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023944, 'name': ReconfigVM_Task, 'duration_secs': 0.549897} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.146275] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Reconfigured VM instance instance-00000033 to attach disk [datastore1] cd0c01ac-602b-44a3-8099-84b8a50b2449/cd0c01ac-602b-44a3-8099-84b8a50b2449.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 749.147048] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f3d074b-be22-4894-b584-951cd5d6f6bb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.155116] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 749.155116] env[63372]: value = "task-1023947" [ 749.155116] env[63372]: _type = "Task" [ 749.155116] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.165620] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023947, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.191410] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023945, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.223118] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.666s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.223662] env[63372]: DEBUG nova.compute.manager [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 749.227147] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.009s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.227529] env[63372]: DEBUG nova.objects.instance [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Lazy-loading 'resources' on Instance uuid c825d2e2-a9cc-4c7a-b92e-039756d4121d {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 749.299472] env[63372]: DEBUG nova.network.neutron [req-2132607c-2657-481d-8ebd-597b8734ee90 req-1573c9a9-e301-458e-86af-492035d4314f service nova] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Updated VIF entry in instance network info cache for port 6c9c3162-8734-4911-a9d1-9f3f9214ea93. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 749.299797] env[63372]: DEBUG nova.network.neutron [req-2132607c-2657-481d-8ebd-597b8734ee90 req-1573c9a9-e301-458e-86af-492035d4314f service nova] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Updating instance_info_cache with network_info: [{"id": "6c9c3162-8734-4911-a9d1-9f3f9214ea93", "address": "fa:16:3e:81:26:4e", "network": {"id": "6a147c66-cd17-47c3-993b-bd3aab62811b", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-786554058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9946a38bddee4f7cb2c6551d76f0a1c0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb3425ea-72e7-41e3-92a7-820db9ec4661", "external-id": "nsx-vlan-transportzone-651", "segmentation_id": 651, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9c3162-87", "ovs_interfaceid": "6c9c3162-8734-4911-a9d1-9f3f9214ea93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.526041] env[63372]: DEBUG nova.compute.manager [req-0337b8a4-25e5-4521-a46f-6679ff36f7b8 req-8e62160a-f65d-4086-a3e2-479ffd9dfbac service nova] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Received event network-vif-plugged-efd54f40-2ca1-4d06-a8c7-428f7508abe9 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.528553] env[63372]: DEBUG oslo_concurrency.lockutils [req-0337b8a4-25e5-4521-a46f-6679ff36f7b8 req-8e62160a-f65d-4086-a3e2-479ffd9dfbac service nova] Acquiring lock "7cf6ac9b-4703-4143-a22c-abc5528ce5ef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.528827] env[63372]: DEBUG oslo_concurrency.lockutils [req-0337b8a4-25e5-4521-a46f-6679ff36f7b8 req-8e62160a-f65d-4086-a3e2-479ffd9dfbac service nova] Lock "7cf6ac9b-4703-4143-a22c-abc5528ce5ef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.529031] env[63372]: DEBUG oslo_concurrency.lockutils [req-0337b8a4-25e5-4521-a46f-6679ff36f7b8 req-8e62160a-f65d-4086-a3e2-479ffd9dfbac service nova] Lock "7cf6ac9b-4703-4143-a22c-abc5528ce5ef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.529226] env[63372]: DEBUG nova.compute.manager [req-0337b8a4-25e5-4521-a46f-6679ff36f7b8 req-8e62160a-f65d-4086-a3e2-479ffd9dfbac service nova] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] No waiting events found dispatching network-vif-plugged-efd54f40-2ca1-4d06-a8c7-428f7508abe9 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 749.529397] env[63372]: WARNING nova.compute.manager [req-0337b8a4-25e5-4521-a46f-6679ff36f7b8 req-8e62160a-f65d-4086-a3e2-479ffd9dfbac service nova] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Received unexpected event network-vif-plugged-efd54f40-2ca1-4d06-a8c7-428f7508abe9 for instance with vm_state building and task_state spawning. [ 749.565545] env[63372]: DEBUG nova.network.neutron [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Successfully updated port: efd54f40-2ca1-4d06-a8c7-428f7508abe9 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 749.600741] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 749.600741] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5276d793-eebe-af70-33f4-9e0adb7191f6" [ 749.600741] env[63372]: _type = "HttpNfcLease" [ 749.600741] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 749.600741] env[63372]: DEBUG oslo_vmware.rw_handles [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 749.600741] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5276d793-eebe-af70-33f4-9e0adb7191f6" [ 749.600741] env[63372]: _type = "HttpNfcLease" [ 749.600741] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 749.601299] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24318b41-3c5f-4983-bff8-0ac36e5f0769 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.609392] env[63372]: DEBUG oslo_vmware.rw_handles [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5292b46d-2875-94c1-da4b-7ef66752c0c7/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 749.609608] env[63372]: DEBUG oslo_vmware.rw_handles [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5292b46d-2875-94c1-da4b-7ef66752c0c7/disk-0.vmdk for reading. {{(pid=63372) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 749.682583] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023947, 'name': Rename_Task, 'duration_secs': 0.45814} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.683480] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 749.686798] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-934ded2e-8050-453d-a66e-5119d28484fe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.697320] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023945, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.797438} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.697320] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] ae6f3f42-7213-4ab1-b74c-1a557df6748b/ae6f3f42-7213-4ab1-b74c-1a557df6748b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 749.697501] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 749.698055] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 749.698055] env[63372]: value = "task-1023948" [ 749.698055] env[63372]: _type = "Task" [ 749.698055] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.698276] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-52c160a5-9b9e-4b22-b5fa-8f0a7ca54323 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.708308] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 749.708308] env[63372]: value = "task-1023949" [ 749.708308] env[63372]: _type = "Task" [ 749.708308] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.712307] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023948, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.722919] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023949, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.733085] env[63372]: DEBUG nova.compute.utils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 749.737727] env[63372]: DEBUG nova.compute.manager [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 749.737992] env[63372]: DEBUG nova.network.neutron [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 749.740282] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b8f2c1c1-77c5-4658-ae84-c16b2d8e1ade {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.802670] env[63372]: DEBUG oslo_concurrency.lockutils [req-2132607c-2657-481d-8ebd-597b8734ee90 req-1573c9a9-e301-458e-86af-492035d4314f service nova] Releasing lock "refresh_cache-7a75a51d-b804-453b-ba7b-fda6c4931802" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.819686] env[63372]: DEBUG nova.policy [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb233d4b6fe54fefae3046d62d167e34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d0c0f315749429dbae22a72ae82b1b2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 750.070471] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "refresh_cache-7cf6ac9b-4703-4143-a22c-abc5528ce5ef" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.070665] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquired lock "refresh_cache-7cf6ac9b-4703-4143-a22c-abc5528ce5ef" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.070819] env[63372]: DEBUG nova.network.neutron [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 750.206810] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fe0b8f-e3db-40b8-8f0a-35895532cbb0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.218221] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023948, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.222492] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4c9ee0-63c1-4772-ad9f-a5a81753b76e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.233162] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023949, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068548} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.260929] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 750.260929] env[63372]: DEBUG nova.compute.manager [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 750.265819] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52caec3b-b893-4640-b226-9c16749a8a79 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.268868] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ad4160-6b77-49f5-ba6a-5fde0e8ad58c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.299441] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] ae6f3f42-7213-4ab1-b74c-1a557df6748b/ae6f3f42-7213-4ab1-b74c-1a557df6748b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 750.300288] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14c97ad6-b3ca-4d00-93c4-7a2eaeeb450a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.319553] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2ea9554-092a-4bcf-85b2-beef4b67fb70 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.337749] env[63372]: DEBUG nova.compute.provider_tree [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.343225] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 750.343225] env[63372]: value = "task-1023950" [ 750.343225] env[63372]: _type = "Task" [ 750.343225] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.350880] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023950, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.389539] env[63372]: DEBUG nova.network.neutron [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Successfully created port: f6d823dd-d794-4292-8906-9988287ccff0 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 750.619200] env[63372]: DEBUG nova.network.neutron [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.693279] env[63372]: DEBUG nova.compute.manager [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 750.694794] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d65409-32e2-41e4-bdac-02e8fc3e553f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.720257] env[63372]: DEBUG oslo_vmware.api [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023948, 'name': PowerOnVM_Task, 'duration_secs': 0.741815} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.720754] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 750.721032] env[63372]: INFO nova.compute.manager [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Took 9.33 seconds to spawn the instance on the hypervisor. [ 750.721353] env[63372]: DEBUG nova.compute.manager [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 750.722303] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9744650-840a-4c5e-84b3-0bba358c4660 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.842834] env[63372]: DEBUG nova.scheduler.client.report [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.857405] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023950, 'name': ReconfigVM_Task, 'duration_secs': 0.293352} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.857889] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Reconfigured VM instance instance-00000034 to attach disk [datastore1] ae6f3f42-7213-4ab1-b74c-1a557df6748b/ae6f3f42-7213-4ab1-b74c-1a557df6748b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 750.858333] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2abfd927-3e46-4562-a03e-f7bc818be681 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.864751] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 750.864751] env[63372]: value = "task-1023951" [ 750.864751] env[63372]: _type = "Task" [ 750.864751] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.880575] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023951, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.911688] env[63372]: DEBUG nova.network.neutron [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Updating instance_info_cache with network_info: [{"id": "efd54f40-2ca1-4d06-a8c7-428f7508abe9", "address": "fa:16:3e:1e:1a:fb", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.182", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapefd54f40-2c", "ovs_interfaceid": "efd54f40-2ca1-4d06-a8c7-428f7508abe9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.082249] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "99f901a6-9bb3-4403-af0c-c8900f655cb3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.082558] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "99f901a6-9bb3-4403-af0c-c8900f655cb3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.209204] env[63372]: INFO nova.compute.manager [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] instance snapshotting [ 751.212395] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c887d4d4-3e52-4926-a9b6-50de73652d3f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.245999] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5199596-2cb1-4849-94eb-94d4396137e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.249597] env[63372]: INFO nova.compute.manager [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Took 37.44 seconds to build instance. [ 751.279278] env[63372]: DEBUG nova.compute.manager [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 751.311976] env[63372]: DEBUG nova.virt.hardware [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 751.312266] env[63372]: DEBUG nova.virt.hardware [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 751.312426] env[63372]: DEBUG nova.virt.hardware [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 751.312684] env[63372]: DEBUG nova.virt.hardware [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 751.312890] env[63372]: DEBUG nova.virt.hardware [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 751.313573] env[63372]: DEBUG nova.virt.hardware [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 751.314574] env[63372]: DEBUG nova.virt.hardware [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 751.314732] env[63372]: DEBUG nova.virt.hardware [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 751.315030] env[63372]: DEBUG nova.virt.hardware [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 751.315245] env[63372]: DEBUG nova.virt.hardware [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 751.315474] env[63372]: DEBUG nova.virt.hardware [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 751.316843] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0e3af0-343b-491f-b5cb-3486f6f863d8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.326049] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c726719c-b7f9-4f5e-a657-afb1123f60ee {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.352862] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.126s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.355870] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 21.989s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.377903] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023951, 'name': Rename_Task, 'duration_secs': 0.148317} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.378972] env[63372]: INFO nova.scheduler.client.report [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Deleted allocations for instance c825d2e2-a9cc-4c7a-b92e-039756d4121d [ 751.380016] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 751.382427] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce08d7cb-e1d9-4a79-a806-4cf64e8aaea0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.392372] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 751.392372] env[63372]: value = "task-1023952" [ 751.392372] env[63372]: _type = "Task" [ 751.392372] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.402289] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023952, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.415078] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Releasing lock "refresh_cache-7cf6ac9b-4703-4143-a22c-abc5528ce5ef" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.415437] env[63372]: DEBUG nova.compute.manager [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Instance network_info: |[{"id": "efd54f40-2ca1-4d06-a8c7-428f7508abe9", "address": "fa:16:3e:1e:1a:fb", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.182", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapefd54f40-2c", "ovs_interfaceid": "efd54f40-2ca1-4d06-a8c7-428f7508abe9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 751.416066] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:1a:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'efd54f40-2ca1-4d06-a8c7-428f7508abe9', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 751.424496] env[63372]: DEBUG oslo.service.loopingcall [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.424765] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 751.425009] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d1ea18aa-0f86-4dae-98cc-9e43d87ddebc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.449981] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 751.449981] env[63372]: value = "task-1023953" [ 751.449981] env[63372]: _type = "Task" [ 751.449981] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.460470] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023953, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.576952] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "de1b38a4-c7f2-420c-a050-7311976e4ca8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.577389] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "de1b38a4-c7f2-420c-a050-7311976e4ca8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.609138] env[63372]: DEBUG nova.compute.manager [req-4f128536-d208-481e-85a0-a2449bc047d1 req-7f70ce0a-8feb-4865-9323-6faced6661bc service nova] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Received event network-changed-efd54f40-2ca1-4d06-a8c7-428f7508abe9 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.609138] env[63372]: DEBUG nova.compute.manager [req-4f128536-d208-481e-85a0-a2449bc047d1 req-7f70ce0a-8feb-4865-9323-6faced6661bc service nova] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Refreshing instance network info cache due to event network-changed-efd54f40-2ca1-4d06-a8c7-428f7508abe9. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 751.609259] env[63372]: DEBUG oslo_concurrency.lockutils [req-4f128536-d208-481e-85a0-a2449bc047d1 req-7f70ce0a-8feb-4865-9323-6faced6661bc service nova] Acquiring lock "refresh_cache-7cf6ac9b-4703-4143-a22c-abc5528ce5ef" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.609317] env[63372]: DEBUG oslo_concurrency.lockutils [req-4f128536-d208-481e-85a0-a2449bc047d1 req-7f70ce0a-8feb-4865-9323-6faced6661bc service nova] Acquired lock "refresh_cache-7cf6ac9b-4703-4143-a22c-abc5528ce5ef" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.609500] env[63372]: DEBUG nova.network.neutron [req-4f128536-d208-481e-85a0-a2449bc047d1 req-7f70ce0a-8feb-4865-9323-6faced6661bc service nova] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Refreshing network info cache for port efd54f40-2ca1-4d06-a8c7-428f7508abe9 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 751.752597] env[63372]: DEBUG oslo_concurrency.lockutils [None req-407096d1-487c-4548-9d6c-c0cd80527ff7 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "cd0c01ac-602b-44a3-8099-84b8a50b2449" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.110s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.759974] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 751.760764] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-33ae2c87-a9c2-46af-91cb-67523aa2f606 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.768107] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 751.768107] env[63372]: value = "task-1023954" [ 751.768107] env[63372]: _type = "Task" [ 751.768107] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.780362] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023954, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.889162] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d77e5f76-5884-49a0-9473-889ac081ee71 tempest-InstanceActionsTestJSON-834543351 tempest-InstanceActionsTestJSON-834543351-project-member] Lock "c825d2e2-a9cc-4c7a-b92e-039756d4121d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.698s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.909399] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023952, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.960861] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023953, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.983372] env[63372]: DEBUG nova.network.neutron [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Successfully updated port: f6d823dd-d794-4292-8906-9988287ccff0 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 752.145024] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "7c9d551e-d210-4943-971f-b2829751fcfc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.145521] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "7c9d551e-d210-4943-971f-b2829751fcfc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.209905] env[63372]: INFO nova.compute.manager [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Rescuing [ 752.210243] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "refresh_cache-cd0c01ac-602b-44a3-8099-84b8a50b2449" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.210413] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquired lock "refresh_cache-cd0c01ac-602b-44a3-8099-84b8a50b2449" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.211632] env[63372]: DEBUG nova.network.neutron [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 752.256405] env[63372]: DEBUG nova.compute.manager [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 752.285645] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023954, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.367793] env[63372]: DEBUG nova.network.neutron [req-4f128536-d208-481e-85a0-a2449bc047d1 req-7f70ce0a-8feb-4865-9323-6faced6661bc service nova] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Updated VIF entry in instance network info cache for port efd54f40-2ca1-4d06-a8c7-428f7508abe9. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 752.368261] env[63372]: DEBUG nova.network.neutron [req-4f128536-d208-481e-85a0-a2449bc047d1 req-7f70ce0a-8feb-4865-9323-6faced6661bc service nova] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Updating instance_info_cache with network_info: [{"id": "efd54f40-2ca1-4d06-a8c7-428f7508abe9", "address": "fa:16:3e:1e:1a:fb", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.182", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapefd54f40-2c", "ovs_interfaceid": "efd54f40-2ca1-4d06-a8c7-428f7508abe9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.377353] env[63372]: INFO nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating resource usage from migration e4ca8d01-990e-4e1b-8d4a-067812994b8f [ 752.400967] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance cf673ac1-2c7d-468b-83ec-c723d5182457 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.401201] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance bbba9f28-045d-41ab-8539-5b2968fe3d54 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.401346] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance ac90a156-be00-4f62-a76e-e08914531167 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.401397] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 5819c38e-2cf9-4d16-b28a-5f23d35c3d44 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.401694] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance c4718797-aa86-4ec0-94d3-6480bd6aa898 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.401891] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.401958] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance d50919ac-3a0b-46ac-a837-ca3e6ad05173 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.402079] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.402402] env[63372]: WARNING nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 63b9095c-fb49-4dc1-a6dc-96529aeaab81 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 752.402559] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance b182294d-2de8-4189-af7f-3e2d2c604a8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.402784] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance f0c60559-c072-4b61-afe8-03d6c131b307 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.402920] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 93a5d948-0629-4f53-a681-858d519acfa7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.403064] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance f43555ef-b517-4b7d-9d2e-4787d40e201b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.403179] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 7a75a51d-b804-453b-ba7b-fda6c4931802 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.403289] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance cd0c01ac-602b-44a3-8099-84b8a50b2449 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.403398] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance ae6f3f42-7213-4ab1-b74c-1a557df6748b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.403507] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 7cf6ac9b-4703-4143-a22c-abc5528ce5ef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.403615] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 757df632-4af2-4eb3-bc6d-8c0812ba850d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 752.409369] env[63372]: DEBUG oslo_vmware.api [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023952, 'name': PowerOnVM_Task, 'duration_secs': 0.533411} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.409840] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 752.410061] env[63372]: INFO nova.compute.manager [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Took 8.42 seconds to spawn the instance on the hypervisor. [ 752.410248] env[63372]: DEBUG nova.compute.manager [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 752.411015] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7074c0-1065-4a31-b0d1-ba488e49ebd7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.466096] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023953, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.492198] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "refresh_cache-757df632-4af2-4eb3-bc6d-8c0812ba850d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.492355] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "refresh_cache-757df632-4af2-4eb3-bc6d-8c0812ba850d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.493164] env[63372]: DEBUG nova.network.neutron [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 752.781638] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023954, 'name': CreateSnapshot_Task, 'duration_secs': 0.885921} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.782731] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.783048] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 752.783861] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f1e104-edb0-4112-9c50-a61623089a35 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.871400] env[63372]: DEBUG oslo_concurrency.lockutils [req-4f128536-d208-481e-85a0-a2449bc047d1 req-7f70ce0a-8feb-4865-9323-6faced6661bc service nova] Releasing lock "refresh_cache-7cf6ac9b-4703-4143-a22c-abc5528ce5ef" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.912250] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 69a107d0-80c7-42e5-b514-b4273e1a3359 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 752.938814] env[63372]: INFO nova.compute.manager [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Took 33.78 seconds to build instance. [ 752.964200] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023953, 'name': CreateVM_Task, 'duration_secs': 1.343827} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.964411] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 752.965249] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.965516] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.965975] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 752.966337] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07615aee-5212-4375-b556-c34cf31c825f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.972337] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 752.972337] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525ea9cb-9e7c-b4ed-822b-f3a26835f04e" [ 752.972337] env[63372]: _type = "Task" [ 752.972337] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.981162] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525ea9cb-9e7c-b4ed-822b-f3a26835f04e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.012233] env[63372]: DEBUG nova.network.neutron [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Updating instance_info_cache with network_info: [{"id": "263cc51f-62e9-4a41-9d93-6f8c90cb320d", "address": "fa:16:3e:80:76:5b", "network": {"id": "2f241a45-cc18-46f6-b0a3-1aad8d2713fd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1953733697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a14a1aa8a0574b65a312754ae1091f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a88b707-352e-4be7-b1d6-ad6074b40ed9", "external-id": "nsx-vlan-transportzone-789", "segmentation_id": 789, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap263cc51f-62", "ovs_interfaceid": "263cc51f-62e9-4a41-9d93-6f8c90cb320d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.026945] env[63372]: DEBUG nova.network.neutron [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 753.247534] env[63372]: DEBUG nova.network.neutron [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Updating instance_info_cache with network_info: [{"id": "f6d823dd-d794-4292-8906-9988287ccff0", "address": "fa:16:3e:d0:a6:9f", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6d823dd-d7", "ovs_interfaceid": "f6d823dd-d794-4292-8906-9988287ccff0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.304870] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 753.304870] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-74ac28ed-cee3-47c5-a728-9395146d5be2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.313825] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 753.313825] env[63372]: value = "task-1023955" [ 753.313825] env[63372]: _type = "Task" [ 753.313825] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.323971] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023955, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.416916] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance febeca1c-3370-4c74-8bd9-efad313df1e2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 753.446358] env[63372]: DEBUG oslo_concurrency.lockutils [None req-75f1b24b-d5ac-4684-ba79-950a1ef57310 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "ae6f3f42-7213-4ab1-b74c-1a557df6748b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.288s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.488697] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525ea9cb-9e7c-b4ed-822b-f3a26835f04e, 'name': SearchDatastore_Task, 'duration_secs': 0.014326} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.488697] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.488697] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 753.488697] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.488697] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.488697] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 753.488697] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ded4df8-9530-42cb-9c34-ae2516dd0986 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.499360] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 753.499360] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 753.500388] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d707963-ff76-4c42-82a4-c535ba055b51 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.506337] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 753.506337] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52bb527e-c0e9-9032-116c-3be161680afd" [ 753.506337] env[63372]: _type = "Task" [ 753.506337] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.516207] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Releasing lock "refresh_cache-cd0c01ac-602b-44a3-8099-84b8a50b2449" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.518636] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52bb527e-c0e9-9032-116c-3be161680afd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.751054] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "refresh_cache-757df632-4af2-4eb3-bc6d-8c0812ba850d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.751493] env[63372]: DEBUG nova.compute.manager [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Instance network_info: |[{"id": "f6d823dd-d794-4292-8906-9988287ccff0", "address": "fa:16:3e:d0:a6:9f", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6d823dd-d7", "ovs_interfaceid": "f6d823dd-d794-4292-8906-9988287ccff0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 753.752035] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:a6:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60badc2d-69d2-467d-a92e-98511f5cb0b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f6d823dd-d794-4292-8906-9988287ccff0', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 753.761947] env[63372]: DEBUG oslo.service.loopingcall [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.762323] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 753.762686] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35705acc-5f05-4186-a432-a9683d69063e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.784306] env[63372]: DEBUG nova.compute.manager [req-a57f7862-9f8a-4e8d-879e-c50c90a8d5a5 req-42616d2b-48ff-4ac9-b0b2-1ed87cebc9c7 service nova] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Received event network-vif-plugged-f6d823dd-d794-4292-8906-9988287ccff0 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.786023] env[63372]: DEBUG oslo_concurrency.lockutils [req-a57f7862-9f8a-4e8d-879e-c50c90a8d5a5 req-42616d2b-48ff-4ac9-b0b2-1ed87cebc9c7 service nova] Acquiring lock "757df632-4af2-4eb3-bc6d-8c0812ba850d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.786023] env[63372]: DEBUG oslo_concurrency.lockutils [req-a57f7862-9f8a-4e8d-879e-c50c90a8d5a5 req-42616d2b-48ff-4ac9-b0b2-1ed87cebc9c7 service nova] Lock "757df632-4af2-4eb3-bc6d-8c0812ba850d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.786023] env[63372]: DEBUG oslo_concurrency.lockutils [req-a57f7862-9f8a-4e8d-879e-c50c90a8d5a5 req-42616d2b-48ff-4ac9-b0b2-1ed87cebc9c7 service nova] Lock "757df632-4af2-4eb3-bc6d-8c0812ba850d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.786023] env[63372]: DEBUG nova.compute.manager [req-a57f7862-9f8a-4e8d-879e-c50c90a8d5a5 req-42616d2b-48ff-4ac9-b0b2-1ed87cebc9c7 service nova] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] No waiting events found dispatching network-vif-plugged-f6d823dd-d794-4292-8906-9988287ccff0 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 753.786023] env[63372]: WARNING nova.compute.manager [req-a57f7862-9f8a-4e8d-879e-c50c90a8d5a5 req-42616d2b-48ff-4ac9-b0b2-1ed87cebc9c7 service nova] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Received unexpected event network-vif-plugged-f6d823dd-d794-4292-8906-9988287ccff0 for instance with vm_state building and task_state spawning. [ 753.786023] env[63372]: DEBUG nova.compute.manager [req-a57f7862-9f8a-4e8d-879e-c50c90a8d5a5 req-42616d2b-48ff-4ac9-b0b2-1ed87cebc9c7 service nova] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Received event network-changed-f6d823dd-d794-4292-8906-9988287ccff0 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.786023] env[63372]: DEBUG nova.compute.manager [req-a57f7862-9f8a-4e8d-879e-c50c90a8d5a5 req-42616d2b-48ff-4ac9-b0b2-1ed87cebc9c7 service nova] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Refreshing instance network info cache due to event network-changed-f6d823dd-d794-4292-8906-9988287ccff0. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 753.786559] env[63372]: DEBUG oslo_concurrency.lockutils [req-a57f7862-9f8a-4e8d-879e-c50c90a8d5a5 req-42616d2b-48ff-4ac9-b0b2-1ed87cebc9c7 service nova] Acquiring lock "refresh_cache-757df632-4af2-4eb3-bc6d-8c0812ba850d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.786559] env[63372]: DEBUG oslo_concurrency.lockutils [req-a57f7862-9f8a-4e8d-879e-c50c90a8d5a5 req-42616d2b-48ff-4ac9-b0b2-1ed87cebc9c7 service nova] Acquired lock "refresh_cache-757df632-4af2-4eb3-bc6d-8c0812ba850d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.786559] env[63372]: DEBUG nova.network.neutron [req-a57f7862-9f8a-4e8d-879e-c50c90a8d5a5 req-42616d2b-48ff-4ac9-b0b2-1ed87cebc9c7 service nova] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Refreshing network info cache for port f6d823dd-d794-4292-8906-9988287ccff0 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 753.790257] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 753.790257] env[63372]: value = "task-1023956" [ 753.790257] env[63372]: _type = "Task" [ 753.790257] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.800945] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023956, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.825657] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023955, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.920368] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 4c9eb955-3bed-4b26-866f-b2a876b835dd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 753.948641] env[63372]: DEBUG nova.compute.manager [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 754.017381] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52bb527e-c0e9-9032-116c-3be161680afd, 'name': SearchDatastore_Task, 'duration_secs': 0.013513} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.018260] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12fc299f-76b5-4382-9edd-afa1ae4c3493 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.024019] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 754.024019] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52340e99-a936-7c39-3e10-147efae36ae1" [ 754.024019] env[63372]: _type = "Task" [ 754.024019] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.032156] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52340e99-a936-7c39-3e10-147efae36ae1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.054614] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 754.054914] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9fbf33e-522f-4074-a97e-35e41a48c462 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.061437] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 754.061437] env[63372]: value = "task-1023957" [ 754.061437] env[63372]: _type = "Task" [ 754.061437] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.070084] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023957, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.301571] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023956, 'name': CreateVM_Task, 'duration_secs': 0.488755} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.301788] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 754.302539] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.302711] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.303076] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 754.303345] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4dce57f-9e8c-4b0e-b6da-e02beb58f212 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.308593] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 754.308593] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]524ee76b-6730-8438-e9fd-bebf98084a58" [ 754.308593] env[63372]: _type = "Task" [ 754.308593] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.316813] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524ee76b-6730-8438-e9fd-bebf98084a58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.324738] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023955, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.334434] env[63372]: DEBUG nova.compute.manager [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 754.335571] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274d1780-bbdb-45cb-9c3c-dc8b04fb21d7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.425300] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance bad91d77-c7ee-4572-b1ed-068b2a55233c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 754.476547] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.537641] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52340e99-a936-7c39-3e10-147efae36ae1, 'name': SearchDatastore_Task, 'duration_secs': 0.027977} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.538043] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.538299] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 7cf6ac9b-4703-4143-a22c-abc5528ce5ef/7cf6ac9b-4703-4143-a22c-abc5528ce5ef.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 754.538617] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aaa2233e-3448-4cdf-81f6-6d16258d4c5f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.545288] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 754.545288] env[63372]: value = "task-1023958" [ 754.545288] env[63372]: _type = "Task" [ 754.545288] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.553900] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023958, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.576580] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023957, 'name': PowerOffVM_Task, 'duration_secs': 0.384136} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.576757] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 754.578139] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b30121-01ec-49a4-89e5-01a26fed36d4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.606338] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6795987d-b9d4-4726-8d39-25133760c501 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.644217] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 754.644567] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-900748c4-283b-440c-b559-1aa305b5074c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.654543] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 754.654543] env[63372]: value = "task-1023959" [ 754.654543] env[63372]: _type = "Task" [ 754.654543] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.667911] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] VM already powered off {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 754.668202] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 754.669020] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.669020] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.669020] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 754.669308] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c04f62d8-3672-492f-a717-7446007dc39c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.680899] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 754.681155] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 754.681967] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-914a37d4-0140-4d65-902c-e566e49e4e16 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.688361] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 754.688361] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]527ab338-321a-982c-98bd-9201304409d8" [ 754.688361] env[63372]: _type = "Task" [ 754.688361] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.697156] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527ab338-321a-982c-98bd-9201304409d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.698228] env[63372]: DEBUG nova.network.neutron [req-a57f7862-9f8a-4e8d-879e-c50c90a8d5a5 req-42616d2b-48ff-4ac9-b0b2-1ed87cebc9c7 service nova] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Updated VIF entry in instance network info cache for port f6d823dd-d794-4292-8906-9988287ccff0. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 754.698595] env[63372]: DEBUG nova.network.neutron [req-a57f7862-9f8a-4e8d-879e-c50c90a8d5a5 req-42616d2b-48ff-4ac9-b0b2-1ed87cebc9c7 service nova] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Updating instance_info_cache with network_info: [{"id": "f6d823dd-d794-4292-8906-9988287ccff0", "address": "fa:16:3e:d0:a6:9f", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6d823dd-d7", "ovs_interfaceid": "f6d823dd-d794-4292-8906-9988287ccff0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.820096] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524ee76b-6730-8438-e9fd-bebf98084a58, 'name': SearchDatastore_Task, 'duration_secs': 0.020861} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.823566] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.823917] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 754.824208] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.830953] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023955, 'name': CloneVM_Task, 'duration_secs': 1.429013} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.831998] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Created linked-clone VM from snapshot [ 754.833265] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c6f26a-2268-49da-bf58-5aecbb79193a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.845382] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Uploading image e429c1e3-9e6d-4a1e-9c01-685c9978c5d9 {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 754.850415] env[63372]: INFO nova.compute.manager [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] instance snapshotting [ 754.853623] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f8a1ea-7805-4387-9be1-088e3ddc80ed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.879281] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e1dfafc-b87a-4580-a3f5-a138f7781fcc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.892292] env[63372]: DEBUG oslo_vmware.rw_handles [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 754.892292] env[63372]: value = "vm-227330" [ 754.892292] env[63372]: _type = "VirtualMachine" [ 754.892292] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 754.892720] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4d11a85d-61a4-4947-9387-90668122d7f0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.903948] env[63372]: DEBUG oslo_vmware.rw_handles [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lease: (returnval){ [ 754.903948] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f723d9-444d-eb28-97e3-74c7563e67f3" [ 754.903948] env[63372]: _type = "HttpNfcLease" [ 754.903948] env[63372]: } obtained for exporting VM: (result){ [ 754.903948] env[63372]: value = "vm-227330" [ 754.903948] env[63372]: _type = "VirtualMachine" [ 754.903948] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 754.904312] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the lease: (returnval){ [ 754.904312] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f723d9-444d-eb28-97e3-74c7563e67f3" [ 754.904312] env[63372]: _type = "HttpNfcLease" [ 754.904312] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 754.911384] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 754.911384] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f723d9-444d-eb28-97e3-74c7563e67f3" [ 754.911384] env[63372]: _type = "HttpNfcLease" [ 754.911384] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 754.928556] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance cbd55518-a3b2-4636-ba43-c279fdd0bf8d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.054898] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023958, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.203028] env[63372]: DEBUG oslo_concurrency.lockutils [req-a57f7862-9f8a-4e8d-879e-c50c90a8d5a5 req-42616d2b-48ff-4ac9-b0b2-1ed87cebc9c7 service nova] Releasing lock "refresh_cache-757df632-4af2-4eb3-bc6d-8c0812ba850d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.204044] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527ab338-321a-982c-98bd-9201304409d8, 'name': SearchDatastore_Task, 'duration_secs': 0.012463} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.204312] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1441fd7c-fe97-414b-8c60-3937208836dd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.210515] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 755.210515] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5299b0c1-cf78-693f-443b-57c8f8f53e72" [ 755.210515] env[63372]: _type = "Task" [ 755.210515] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.219268] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5299b0c1-cf78-693f-443b-57c8f8f53e72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.400047] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 755.400334] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-57b065f4-6fcc-465a-8af9-65415bbcfd3f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.409202] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 755.409202] env[63372]: value = "task-1023961" [ 755.409202] env[63372]: _type = "Task" [ 755.409202] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.413642] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 755.413642] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f723d9-444d-eb28-97e3-74c7563e67f3" [ 755.413642] env[63372]: _type = "HttpNfcLease" [ 755.413642] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 755.414261] env[63372]: DEBUG oslo_vmware.rw_handles [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 755.414261] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f723d9-444d-eb28-97e3-74c7563e67f3" [ 755.414261] env[63372]: _type = "HttpNfcLease" [ 755.414261] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 755.415953] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a252c04b-4a55-47d0-ad0f-520d63108c1e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.420430] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023961, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.425387] env[63372]: DEBUG oslo_vmware.rw_handles [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5285e3cb-256f-ba12-fd9a-614090c1e927/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 755.425387] env[63372]: DEBUG oslo_vmware.rw_handles [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5285e3cb-256f-ba12-fd9a-614090c1e927/disk-0.vmdk for reading. {{(pid=63372) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 755.484539] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 3f66eda6-f5e9-4527-9711-849a01702580 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.558021] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023958, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.995619} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.558021] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 7cf6ac9b-4703-4143-a22c-abc5528ce5ef/7cf6ac9b-4703-4143-a22c-abc5528ce5ef.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 755.558248] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 755.558404] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75a0c7c5-b8ef-473d-9f84-3953d3b47de0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.566690] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-adaa5edc-11c5-4c25-9f1a-13f6e71abf5a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.569065] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 755.569065] env[63372]: value = "task-1023962" [ 755.569065] env[63372]: _type = "Task" [ 755.569065] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.579378] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023962, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.721944] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5299b0c1-cf78-693f-443b-57c8f8f53e72, 'name': SearchDatastore_Task, 'duration_secs': 0.014412} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.722277] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.722552] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] cd0c01ac-602b-44a3-8099-84b8a50b2449/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk. {{(pid=63372) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 755.722850] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.723059] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 755.723291] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0039f815-5d61-49cb-ba7d-ec0bce41d8b2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.725722] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f785211-a7a9-4cef-8fc5-7a25b597b466 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.732288] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 755.732288] env[63372]: value = "task-1023963" [ 755.732288] env[63372]: _type = "Task" [ 755.732288] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.736614] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 755.736614] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 755.737729] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f25548e-0078-4d3b-972a-cd5f76b28ffe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.743045] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023963, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.746134] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 755.746134] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52208ef2-6c6d-b059-18d4-7184c07a498e" [ 755.746134] env[63372]: _type = "Task" [ 755.746134] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.753684] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52208ef2-6c6d-b059-18d4-7184c07a498e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.919370] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023961, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.987437] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance f6675697-1529-46be-b28a-398ff3060d18 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.987733] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Migration e4ca8d01-990e-4e1b-8d4a-067812994b8f is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 755.987733] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 8283b736-ad02-4082-97b7-561bd5c5da93 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 756.094691] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023962, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.322003} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.099038] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 756.099038] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c2f0dd-41fd-4746-a978-304b24db98c3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.131548] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 7cf6ac9b-4703-4143-a22c-abc5528ce5ef/7cf6ac9b-4703-4143-a22c-abc5528ce5ef.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 756.132035] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69d0531e-1862-4f1c-bee1-90d2c9726c36 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.160671] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 756.160671] env[63372]: value = "task-1023964" [ 756.160671] env[63372]: _type = "Task" [ 756.160671] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.170409] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023964, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.251618] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023963, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.265700] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52208ef2-6c6d-b059-18d4-7184c07a498e, 'name': SearchDatastore_Task, 'duration_secs': 0.0112} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.267066] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0aa24396-abf8-4b06-a47a-2db85573d273 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.274654] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 756.274654] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525b0937-4d03-0381-2cf6-2791ae748f6a" [ 756.274654] env[63372]: _type = "Task" [ 756.274654] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.286258] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525b0937-4d03-0381-2cf6-2791ae748f6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.423407] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023961, 'name': CreateSnapshot_Task, 'duration_secs': 0.825785} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.423407] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 756.427497] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4b159c-e7f6-4a82-a400-44785ae4b653 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.493459] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 99f901a6-9bb3-4403-af0c-c8900f655cb3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.672252] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023964, 'name': ReconfigVM_Task, 'duration_secs': 0.35828} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.672870] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 7cf6ac9b-4703-4143-a22c-abc5528ce5ef/7cf6ac9b-4703-4143-a22c-abc5528ce5ef.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 756.673385] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93dd4e6c-8043-4b8c-8247-3d7835e9f30e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.681526] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 756.681526] env[63372]: value = "task-1023965" [ 756.681526] env[63372]: _type = "Task" [ 756.681526] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.691182] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023965, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.743518] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023963, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.620868} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.744553] env[63372]: INFO nova.virt.vmwareapi.ds_util [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] cd0c01ac-602b-44a3-8099-84b8a50b2449/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk. [ 756.744714] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d68bf282-dfd7-4347-a0d1-d7c8a731bcdf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.772899] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] cd0c01ac-602b-44a3-8099-84b8a50b2449/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 756.773368] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b909b819-6764-465e-8b3d-e754f6f2d716 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.801533] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525b0937-4d03-0381-2cf6-2791ae748f6a, 'name': SearchDatastore_Task, 'duration_secs': 0.052303} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.802455] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.802859] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 757df632-4af2-4eb3-bc6d-8c0812ba850d/757df632-4af2-4eb3-bc6d-8c0812ba850d.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 756.803375] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 756.803375] env[63372]: value = "task-1023966" [ 756.803375] env[63372]: _type = "Task" [ 756.803375] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.803614] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a75ad803-b16b-4a19-8c33-d6b2fba1d271 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.819414] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023966, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.821227] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 756.821227] env[63372]: value = "task-1023967" [ 756.821227] env[63372]: _type = "Task" [ 756.821227] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.831513] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023967, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.951235] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 756.951689] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f27d2b00-ed13-48da-9550-ffb2dc54af9e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.960092] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 756.960092] env[63372]: value = "task-1023968" [ 756.960092] env[63372]: _type = "Task" [ 756.960092] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.969960] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023968, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.994114] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance de1b38a4-c7f2-420c-a050-7311976e4ca8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.192935] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023965, 'name': Rename_Task, 'duration_secs': 0.154865} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.193747] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 757.193879] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2bd0f769-e131-4dc4-8fe5-1b4d6c42a24c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.201431] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 757.201431] env[63372]: value = "task-1023969" [ 757.201431] env[63372]: _type = "Task" [ 757.201431] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.210763] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023969, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.316859] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023966, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.331412] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023967, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.469927] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023968, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.498456] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 7c9d551e-d210-4943-971f-b2829751fcfc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.498456] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Total usable vcpus: 48, total allocated vcpus: 18 {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 757.499084] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3968MB phys_disk=200GB used_disk=17GB total_vcpus=48 used_vcpus=18 pci_stats=[] {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 757.714317] env[63372]: DEBUG oslo_vmware.api [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023969, 'name': PowerOnVM_Task, 'duration_secs': 0.505526} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.714941] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 757.715242] env[63372]: INFO nova.compute.manager [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Took 9.10 seconds to spawn the instance on the hypervisor. [ 757.715727] env[63372]: DEBUG nova.compute.manager [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 757.716781] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7967f7c5-3cc0-4acd-8409-8ceea1bd3507 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.816938] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023966, 'name': ReconfigVM_Task, 'duration_secs': 0.771207} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.820014] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Reconfigured VM instance instance-00000033 to attach disk [datastore1] cd0c01ac-602b-44a3-8099-84b8a50b2449/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 757.821288] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84171d72-59c3-4166-adca-de4a5da8a7ae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.857583] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-869b9cbb-39df-450a-abdd-b84634a28f09 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.868215] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023967, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.632302} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.869817] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 757df632-4af2-4eb3-bc6d-8c0812ba850d/757df632-4af2-4eb3-bc6d-8c0812ba850d.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 757.869817] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 757.870292] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a10048b9-d6d2-47f7-9a47-1afea09fd5dd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.875639] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 757.875639] env[63372]: value = "task-1023970" [ 757.875639] env[63372]: _type = "Task" [ 757.875639] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.881745] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 757.881745] env[63372]: value = "task-1023971" [ 757.881745] env[63372]: _type = "Task" [ 757.881745] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.889349] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023970, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.898461] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023971, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.974629] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023968, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.989075] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49eb2818-939e-4390-b7a7-fbcae9c77799 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.996620] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ece228a-528a-4ba9-a375-164581a54943 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.027443] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8c066e-b31c-4a7d-8f3c-07805a1b94c6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.035398] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a0648e-8e39-4aab-97f3-01b06b28568e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.049703] env[63372]: DEBUG nova.compute.provider_tree [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.244025] env[63372]: INFO nova.compute.manager [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Took 32.72 seconds to build instance. [ 758.397747] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023970, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.406324] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023971, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080186} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.406612] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 758.407508] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2cbfeda-18b1-4810-978a-067fba9fe318 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.429730] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 757df632-4af2-4eb3-bc6d-8c0812ba850d/757df632-4af2-4eb3-bc6d-8c0812ba850d.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 758.430129] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad2a7080-a22d-463d-a5e1-c61feca99e34 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.450601] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 758.450601] env[63372]: value = "task-1023972" [ 758.450601] env[63372]: _type = "Task" [ 758.450601] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.459437] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023972, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.477480] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023968, 'name': CloneVM_Task} progress is 95%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.552669] env[63372]: DEBUG nova.scheduler.client.report [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.746795] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a865836d-c051-4d9b-b1a7-11b681a7be73 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "7cf6ac9b-4703-4143-a22c-abc5528ce5ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.728s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.885678] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023970, 'name': ReconfigVM_Task, 'duration_secs': 0.824786} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.885997] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 758.886270] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6d49617-718c-4ccf-b372-145b374d99e8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.892950] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 758.892950] env[63372]: value = "task-1023973" [ 758.892950] env[63372]: _type = "Task" [ 758.892950] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.901069] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023973, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.962036] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023972, 'name': ReconfigVM_Task, 'duration_secs': 0.429481} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.962036] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 757df632-4af2-4eb3-bc6d-8c0812ba850d/757df632-4af2-4eb3-bc6d-8c0812ba850d.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 758.962328] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-40d6bea2-0e65-4812-bddc-449703418ab4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.969919] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 758.969919] env[63372]: value = "task-1023974" [ 758.969919] env[63372]: _type = "Task" [ 758.969919] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.977862] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023968, 'name': CloneVM_Task, 'duration_secs': 1.778655} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.978238] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Created linked-clone VM from snapshot [ 758.978995] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25877dd6-c621-429d-b38a-d7887cff12d2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.984789] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023974, 'name': Rename_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.990696] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Uploading image 3c9be794-7dcb-43df-a9b2-d9289ac8889f {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 759.059078] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 759.059396] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.704s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.059645] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.674s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.061314] env[63372]: INFO nova.compute.claims [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.177386] env[63372]: DEBUG oslo_vmware.rw_handles [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 759.177386] env[63372]: value = "vm-227333" [ 759.177386] env[63372]: _type = "VirtualMachine" [ 759.177386] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 759.180185] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9057a8bc-9564-4d8e-8326-090cc0d53c91 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.189022] env[63372]: DEBUG oslo_vmware.rw_handles [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5292b46d-2875-94c1-da4b-7ef66752c0c7/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 759.189938] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef828df-2b15-4b2e-9bc6-75ceaebda03e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.195733] env[63372]: DEBUG oslo_vmware.rw_handles [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lease: (returnval){ [ 759.195733] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]528bf723-9d8e-186b-ddbe-90f2a4f9b5d8" [ 759.195733] env[63372]: _type = "HttpNfcLease" [ 759.195733] env[63372]: } obtained for exporting VM: (result){ [ 759.195733] env[63372]: value = "vm-227333" [ 759.195733] env[63372]: _type = "VirtualMachine" [ 759.195733] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 759.196182] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the lease: (returnval){ [ 759.196182] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]528bf723-9d8e-186b-ddbe-90f2a4f9b5d8" [ 759.196182] env[63372]: _type = "HttpNfcLease" [ 759.196182] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 759.201438] env[63372]: DEBUG oslo_vmware.rw_handles [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5292b46d-2875-94c1-da4b-7ef66752c0c7/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 759.201619] env[63372]: ERROR oslo_vmware.rw_handles [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5292b46d-2875-94c1-da4b-7ef66752c0c7/disk-0.vmdk due to incomplete transfer. [ 759.202275] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3d36703d-0543-4387-affc-621e597663d3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.205401] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 759.205401] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]528bf723-9d8e-186b-ddbe-90f2a4f9b5d8" [ 759.205401] env[63372]: _type = "HttpNfcLease" [ 759.205401] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 759.206062] env[63372]: DEBUG oslo_vmware.rw_handles [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 759.206062] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]528bf723-9d8e-186b-ddbe-90f2a4f9b5d8" [ 759.206062] env[63372]: _type = "HttpNfcLease" [ 759.206062] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 759.206815] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b107f4aa-85fc-4c75-bbb7-1a70cb0a75b2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.211218] env[63372]: DEBUG oslo_vmware.rw_handles [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5292b46d-2875-94c1-da4b-7ef66752c0c7/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 759.211365] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Uploaded image 7dfc6004-7365-4eea-b15a-80afc3639b32 to the Glance image server {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 759.213690] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 759.217317] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6d0da324-eb7b-419d-8d5a-dd26d105e127 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.219074] env[63372]: DEBUG oslo_vmware.rw_handles [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5205f68b-4c16-0d99-acd7-8b608ec1b4e1/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 759.219290] env[63372]: DEBUG oslo_vmware.rw_handles [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5205f68b-4c16-0d99-acd7-8b608ec1b4e1/disk-0.vmdk for reading. {{(pid=63372) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 759.276925] env[63372]: DEBUG nova.compute.manager [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 759.283230] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 759.283230] env[63372]: value = "task-1023976" [ 759.283230] env[63372]: _type = "Task" [ 759.283230] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.293505] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023976, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.359868] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3cfe6a10-f947-4fda-9fa1-dbe562282a94 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.405284] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023973, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.480230] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023974, 'name': Rename_Task, 'duration_secs': 0.201095} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.480511] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 759.480762] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-59860939-6fb3-45c9-9198-4221c8aa23a0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.487503] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 759.487503] env[63372]: value = "task-1023977" [ 759.487503] env[63372]: _type = "Task" [ 759.487503] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.495286] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023977, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.798210] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023976, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.813706] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.906108] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023973, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.998856] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023977, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.301231] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023976, 'name': Destroy_Task, 'duration_secs': 1.00444} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.301231] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Destroyed the VM [ 760.301379] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 760.305042] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-722388e4-38ca-4c78-8c89-f16cde8e98bc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.312857] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 760.312857] env[63372]: value = "task-1023978" [ 760.312857] env[63372]: _type = "Task" [ 760.312857] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.322844] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023978, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.410614] env[63372]: DEBUG oslo_vmware.api [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1023973, 'name': PowerOnVM_Task, 'duration_secs': 1.338525} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.410937] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 760.414264] env[63372]: DEBUG nova.compute.manager [None req-3d5663f5-24a0-4aa0-9fb9-baf6b3270ef0 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 760.415344] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab1e60c3-f853-486b-8e81-76547f15b3d3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.474085] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf06489d-07c7-45ac-8731-3c8816d40be7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.492025] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-08fb7dac-59d6-4255-9fcc-865b873ad5b4 tempest-ServersAdminNegativeTestJSON-107306612 tempest-ServersAdminNegativeTestJSON-107306612-project-admin] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Suspending the VM {{(pid=63372) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 760.492025] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-d09a47db-f2c7-4aac-a8dd-77d855e36170 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.506530] env[63372]: DEBUG oslo_vmware.api [None req-08fb7dac-59d6-4255-9fcc-865b873ad5b4 tempest-ServersAdminNegativeTestJSON-107306612 tempest-ServersAdminNegativeTestJSON-107306612-project-admin] Waiting for the task: (returnval){ [ 760.506530] env[63372]: value = "task-1023979" [ 760.506530] env[63372]: _type = "Task" [ 760.506530] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.506836] env[63372]: DEBUG oslo_vmware.api [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023977, 'name': PowerOnVM_Task, 'duration_secs': 0.627933} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.507199] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 760.507427] env[63372]: INFO nova.compute.manager [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Took 9.23 seconds to spawn the instance on the hypervisor. [ 760.507998] env[63372]: DEBUG nova.compute.manager [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 760.512689] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f89dcb-3a2d-4042-b07c-34e2fb3fb1bc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.530255] env[63372]: DEBUG oslo_vmware.api [None req-08fb7dac-59d6-4255-9fcc-865b873ad5b4 tempest-ServersAdminNegativeTestJSON-107306612 tempest-ServersAdminNegativeTestJSON-107306612-project-admin] Task: {'id': task-1023979, 'name': SuspendVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.635180] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f585de82-eab2-41b7-b438-c7c6e980ae66 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.644810] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0e37c3-ddb3-464b-b799-2eca6b93dca9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.679789] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1dfe72-3958-4cea-bc8a-222645ed7ac9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.689256] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75e8e088-5b94-44d2-a764-9b217add8ae6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.706102] env[63372]: DEBUG nova.compute.provider_tree [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.824225] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023978, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.019977] env[63372]: DEBUG oslo_vmware.api [None req-08fb7dac-59d6-4255-9fcc-865b873ad5b4 tempest-ServersAdminNegativeTestJSON-107306612 tempest-ServersAdminNegativeTestJSON-107306612-project-admin] Task: {'id': task-1023979, 'name': SuspendVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.047115] env[63372]: INFO nova.compute.manager [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Took 33.97 seconds to build instance. [ 761.209156] env[63372]: DEBUG nova.scheduler.client.report [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.324143] env[63372]: DEBUG oslo_vmware.api [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023978, 'name': RemoveSnapshot_Task, 'duration_secs': 0.957939} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.324568] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 761.324900] env[63372]: INFO nova.compute.manager [None req-e1b0ab86-8a14-494f-ab36-838a217023d2 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Took 17.40 seconds to snapshot the instance on the hypervisor. [ 761.520036] env[63372]: DEBUG oslo_vmware.api [None req-08fb7dac-59d6-4255-9fcc-865b873ad5b4 tempest-ServersAdminNegativeTestJSON-107306612 tempest-ServersAdminNegativeTestJSON-107306612-project-admin] Task: {'id': task-1023979, 'name': SuspendVM_Task, 'duration_secs': 0.669169} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.520036] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-08fb7dac-59d6-4255-9fcc-865b873ad5b4 tempest-ServersAdminNegativeTestJSON-107306612 tempest-ServersAdminNegativeTestJSON-107306612-project-admin] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Suspended the VM {{(pid=63372) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 761.520036] env[63372]: DEBUG nova.compute.manager [None req-08fb7dac-59d6-4255-9fcc-865b873ad5b4 tempest-ServersAdminNegativeTestJSON-107306612 tempest-ServersAdminNegativeTestJSON-107306612-project-admin] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 761.520450] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081b6832-ff31-43ed-b469-242b5c0ec1b2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.549808] env[63372]: DEBUG oslo_concurrency.lockutils [None req-29816829-502c-4560-8cdd-b23798895c79 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "757df632-4af2-4eb3-bc6d-8c0812ba850d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.257s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.714405] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.655s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.714879] env[63372]: DEBUG nova.compute.manager [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 761.717904] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.897s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.720263] env[63372]: INFO nova.compute.claims [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.052221] env[63372]: DEBUG nova.compute.manager [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 762.225472] env[63372]: DEBUG nova.compute.utils [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 762.229305] env[63372]: DEBUG nova.compute.manager [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 762.229498] env[63372]: DEBUG nova.network.neutron [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 762.281940] env[63372]: DEBUG nova.policy [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5625f896be74448981e41d104e7ece16', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '541cfc7f0fba40678e1f67d677d22df2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 762.576286] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.607600] env[63372]: DEBUG nova.network.neutron [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Successfully created port: 29995495-a43f-4be1-b907-c88d9fe78dcc {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 762.730457] env[63372]: DEBUG nova.compute.manager [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 763.162527] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23155efb-ef42-4e7f-8a25-cd2028c38221 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.171289] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7cec3d-1491-45bd-869f-a4bf069d5a04 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.204420] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756223e0-b89b-49b0-a4b5-4f4857c2e41b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.212373] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7919ce2-d545-474d-b63f-2fdb755eddfa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.226463] env[63372]: DEBUG nova.compute.provider_tree [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.240702] env[63372]: INFO nova.virt.block_device [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Booting with volume 3226fd91-2732-41ce-bb8e-f909a830c8d4 at /dev/sda [ 763.281000] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e1feb1c8-12a9-41d1-bd5d-837bfc47dbf4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.292233] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861b0d25-d23c-4be9-ae40-b95221ad77bb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.327478] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2549e417-5bcc-4062-b199-d86824b558d7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.336156] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e86dc60-9616-48e2-89c4-734bed697ebe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.373727] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50128bfc-aef4-4fd9-a3b5-aa2ac6b74de1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.381586] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e67ed3-8cfb-4e77-b99b-f11e28187791 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.396574] env[63372]: DEBUG nova.virt.block_device [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Updating existing volume attachment record: 89bf9d4f-2dde-4f84-919d-6e734820561e {{(pid=63372) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 763.730083] env[63372]: DEBUG nova.scheduler.client.report [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.146559] env[63372]: DEBUG nova.network.neutron [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Successfully updated port: 29995495-a43f-4be1-b907-c88d9fe78dcc {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 764.235750] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.236322] env[63372]: DEBUG nova.compute.manager [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 764.239351] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.043s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.240982] env[63372]: INFO nova.compute.claims [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 764.648833] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Acquiring lock "refresh_cache-69a107d0-80c7-42e5-b514-b4273e1a3359" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.649065] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Acquired lock "refresh_cache-69a107d0-80c7-42e5-b514-b4273e1a3359" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.649185] env[63372]: DEBUG nova.network.neutron [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 764.748034] env[63372]: DEBUG nova.compute.utils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 764.750043] env[63372]: DEBUG nova.compute.manager [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 764.750269] env[63372]: DEBUG nova.network.neutron [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 764.794502] env[63372]: DEBUG nova.policy [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2124bf74e1944c7ac4e22dc5fea8bc3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '056172ef521c4de08b5033057773b2ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.062779] env[63372]: DEBUG nova.network.neutron [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Successfully created port: 485f2c37-5aa1-470e-8b39-a986ebd7fd24 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.185366] env[63372]: DEBUG nova.network.neutron [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 765.254708] env[63372]: DEBUG nova.compute.manager [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 765.335824] env[63372]: DEBUG nova.network.neutron [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Successfully created port: 9249f3d1-43e4-4e00-94a1-d4ed0b20b171 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.363630] env[63372]: DEBUG nova.network.neutron [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Updating instance_info_cache with network_info: [{"id": "29995495-a43f-4be1-b907-c88d9fe78dcc", "address": "fa:16:3e:cb:ee:42", "network": {"id": "3ca584f0-a422-4ca2-bdc9-033462b8580e", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-175791853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "541cfc7f0fba40678e1f67d677d22df2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c6a4836-66dc-4e43-982b-f8fcd3f9989a", "external-id": "nsx-vlan-transportzone-635", "segmentation_id": 635, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29995495-a4", "ovs_interfaceid": "29995495-a43f-4be1-b907-c88d9fe78dcc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.491912] env[63372]: DEBUG nova.compute.manager [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 765.492510] env[63372]: DEBUG nova.virt.hardware [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 765.493075] env[63372]: DEBUG nova.virt.hardware [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 765.493075] env[63372]: DEBUG nova.virt.hardware [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 765.493075] env[63372]: DEBUG nova.virt.hardware [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 765.493274] env[63372]: DEBUG nova.virt.hardware [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 765.493341] env[63372]: DEBUG nova.virt.hardware [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 765.493575] env[63372]: DEBUG nova.virt.hardware [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 765.493804] env[63372]: DEBUG nova.virt.hardware [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 765.493998] env[63372]: DEBUG nova.virt.hardware [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 765.494196] env[63372]: DEBUG nova.virt.hardware [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 765.494376] env[63372]: DEBUG nova.virt.hardware [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 765.497724] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f7d0dc0-26eb-4c89-a959-ea896a4b32db {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.506326] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f122dbf-d356-49f5-a1b3-475bb30964f2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.680911] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42786337-a4a8-4a59-9dba-9884564ce8f3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.689012] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd95747-e1a4-4f00-849a-17c996e388d2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.719565] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4353772-3aad-4d16-ad7d-26fc70f760b5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.728463] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a074532-7939-4a89-a71a-7dcbd1b30399 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.746305] env[63372]: DEBUG nova.compute.provider_tree [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.866825] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Releasing lock "refresh_cache-69a107d0-80c7-42e5-b514-b4273e1a3359" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.867260] env[63372]: DEBUG nova.compute.manager [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Instance network_info: |[{"id": "29995495-a43f-4be1-b907-c88d9fe78dcc", "address": "fa:16:3e:cb:ee:42", "network": {"id": "3ca584f0-a422-4ca2-bdc9-033462b8580e", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-175791853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "541cfc7f0fba40678e1f67d677d22df2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c6a4836-66dc-4e43-982b-f8fcd3f9989a", "external-id": "nsx-vlan-transportzone-635", "segmentation_id": 635, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29995495-a4", "ovs_interfaceid": "29995495-a43f-4be1-b907-c88d9fe78dcc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 765.867784] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:ee:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c6a4836-66dc-4e43-982b-f8fcd3f9989a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '29995495-a43f-4be1-b907-c88d9fe78dcc', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 765.875843] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Creating folder: Project (541cfc7f0fba40678e1f67d677d22df2). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 765.876216] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05d5aaeb-abe9-4ede-96cf-3508f63b73e5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.899298] env[63372]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 765.899488] env[63372]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63372) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 765.899849] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Folder already exists: Project (541cfc7f0fba40678e1f67d677d22df2). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 765.900059] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Creating folder: Instances. Parent ref: group-v227264. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 765.900308] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a6a1887d-6e22-4888-8d0e-d11dc1085a7a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.910127] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Created folder: Instances in parent group-v227264. [ 765.910357] env[63372]: DEBUG oslo.service.loopingcall [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.910557] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 765.910772] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d90fbae-55d2-4f34-934c-aeabd61d5683 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.931551] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 765.931551] env[63372]: value = "task-1023982" [ 765.931551] env[63372]: _type = "Task" [ 765.931551] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.944184] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023982, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.250619] env[63372]: DEBUG nova.scheduler.client.report [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.270276] env[63372]: DEBUG nova.compute.manager [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 766.365295] env[63372]: DEBUG nova.virt.hardware [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 766.365295] env[63372]: DEBUG nova.virt.hardware [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 766.365295] env[63372]: DEBUG nova.virt.hardware [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.365693] env[63372]: DEBUG nova.virt.hardware [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 766.365693] env[63372]: DEBUG nova.virt.hardware [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.365693] env[63372]: DEBUG nova.virt.hardware [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 766.365880] env[63372]: DEBUG nova.virt.hardware [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 766.366084] env[63372]: DEBUG nova.virt.hardware [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 766.366228] env[63372]: DEBUG nova.virt.hardware [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 766.366905] env[63372]: DEBUG nova.virt.hardware [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 766.366905] env[63372]: DEBUG nova.virt.hardware [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.369116] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0771056-1ab2-4aae-907d-ed314d3c22f0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.378818] env[63372]: DEBUG oslo_vmware.rw_handles [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5285e3cb-256f-ba12-fd9a-614090c1e927/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 766.380118] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c81ba40-6d85-496c-9a73-d96fcac6f009 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.385861] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d3aa2d-6504-4fbf-a07d-adeb74904045 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.398800] env[63372]: DEBUG oslo_vmware.rw_handles [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5285e3cb-256f-ba12-fd9a-614090c1e927/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 766.398993] env[63372]: ERROR oslo_vmware.rw_handles [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5285e3cb-256f-ba12-fd9a-614090c1e927/disk-0.vmdk due to incomplete transfer. [ 766.399465] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ec875149-f15f-486e-9f40-03764ac5dfd0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.407942] env[63372]: DEBUG oslo_vmware.rw_handles [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5285e3cb-256f-ba12-fd9a-614090c1e927/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 766.408304] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Uploaded image e429c1e3-9e6d-4a1e-9c01-685c9978c5d9 to the Glance image server {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 766.409832] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 766.410561] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-41efb16a-8c21-45de-8cf3-8a669662e6a8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.418214] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 766.418214] env[63372]: value = "task-1023983" [ 766.418214] env[63372]: _type = "Task" [ 766.418214] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.427388] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023983, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.442455] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023982, 'name': CreateVM_Task, 'duration_secs': 0.397302} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.442455] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 766.443169] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': '89bf9d4f-2dde-4f84-919d-6e734820561e', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227276', 'volume_id': '3226fd91-2732-41ce-bb8e-f909a830c8d4', 'name': 'volume-3226fd91-2732-41ce-bb8e-f909a830c8d4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '69a107d0-80c7-42e5-b514-b4273e1a3359', 'attached_at': '', 'detached_at': '', 'volume_id': '3226fd91-2732-41ce-bb8e-f909a830c8d4', 'serial': '3226fd91-2732-41ce-bb8e-f909a830c8d4'}, 'boot_index': 0, 'disk_bus': None, 'device_type': None, 'mount_device': '/dev/sda', 'guest_format': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=63372) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 766.443441] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Root volume attach. Driver type: vmdk {{(pid=63372) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 766.444251] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f888a677-f8e7-45d1-b061-8d34d98aa1d4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.452726] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a9bdc3-37ef-4147-8121-a9efb04c6116 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.459508] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcaa152d-f9cd-48f6-930c-3c340d9269fd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.466551] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-7b159669-c920-4dbe-985a-019b5c43b699 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.473671] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Waiting for the task: (returnval){ [ 766.473671] env[63372]: value = "task-1023984" [ 766.473671] env[63372]: _type = "Task" [ 766.473671] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.481811] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023984, 'name': RelocateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.757577] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.517s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.757577] env[63372]: DEBUG nova.compute.manager [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 766.760176] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 30.529s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.929722] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023983, 'name': Destroy_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.985741] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023984, 'name': RelocateVM_Task} progress is 13%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.265461] env[63372]: DEBUG nova.compute.utils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 767.269480] env[63372]: INFO nova.compute.claims [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 767.273751] env[63372]: DEBUG nova.compute.manager [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 767.274080] env[63372]: DEBUG nova.network.neutron [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 767.332792] env[63372]: DEBUG nova.policy [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5bfa37f6821e426da8b1819a6dd790df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd9532ba6f40b4592bd29d2f78039377a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 767.433761] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023983, 'name': Destroy_Task, 'duration_secs': 0.792764} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.434101] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Destroyed the VM [ 767.434355] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 767.434630] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1a371fd3-0ece-4173-86ec-ead771b6572e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.448510] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 767.448510] env[63372]: value = "task-1023985" [ 767.448510] env[63372]: _type = "Task" [ 767.448510] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.463023] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023985, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.487167] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023984, 'name': RelocateVM_Task} progress is 13%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.490425] env[63372]: DEBUG nova.compute.manager [req-45fdda32-2e91-4107-b6fa-b396917aed9b req-afa0967d-a89a-439e-933d-41a2b5a240e0 service nova] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Received event network-vif-plugged-29995495-a43f-4be1-b907-c88d9fe78dcc {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 767.490749] env[63372]: DEBUG oslo_concurrency.lockutils [req-45fdda32-2e91-4107-b6fa-b396917aed9b req-afa0967d-a89a-439e-933d-41a2b5a240e0 service nova] Acquiring lock "69a107d0-80c7-42e5-b514-b4273e1a3359-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.491206] env[63372]: DEBUG oslo_concurrency.lockutils [req-45fdda32-2e91-4107-b6fa-b396917aed9b req-afa0967d-a89a-439e-933d-41a2b5a240e0 service nova] Lock "69a107d0-80c7-42e5-b514-b4273e1a3359-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.491826] env[63372]: DEBUG oslo_concurrency.lockutils [req-45fdda32-2e91-4107-b6fa-b396917aed9b req-afa0967d-a89a-439e-933d-41a2b5a240e0 service nova] Lock "69a107d0-80c7-42e5-b514-b4273e1a3359-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.492124] env[63372]: DEBUG nova.compute.manager [req-45fdda32-2e91-4107-b6fa-b396917aed9b req-afa0967d-a89a-439e-933d-41a2b5a240e0 service nova] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] No waiting events found dispatching network-vif-plugged-29995495-a43f-4be1-b907-c88d9fe78dcc {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 767.492386] env[63372]: WARNING nova.compute.manager [req-45fdda32-2e91-4107-b6fa-b396917aed9b req-afa0967d-a89a-439e-933d-41a2b5a240e0 service nova] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Received unexpected event network-vif-plugged-29995495-a43f-4be1-b907-c88d9fe78dcc for instance with vm_state building and task_state spawning. [ 767.621454] env[63372]: DEBUG nova.network.neutron [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Successfully updated port: 485f2c37-5aa1-470e-8b39-a986ebd7fd24 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 767.669075] env[63372]: DEBUG nova.network.neutron [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Successfully created port: f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 767.777877] env[63372]: DEBUG nova.compute.manager [req-6a718858-44a6-4b1a-863e-a57fd2c4ea60 req-6894fd71-33e3-4145-9fd5-026219c83b0f service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Received event network-vif-plugged-485f2c37-5aa1-470e-8b39-a986ebd7fd24 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 767.778410] env[63372]: DEBUG oslo_concurrency.lockutils [req-6a718858-44a6-4b1a-863e-a57fd2c4ea60 req-6894fd71-33e3-4145-9fd5-026219c83b0f service nova] Acquiring lock "febeca1c-3370-4c74-8bd9-efad313df1e2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.778410] env[63372]: DEBUG oslo_concurrency.lockutils [req-6a718858-44a6-4b1a-863e-a57fd2c4ea60 req-6894fd71-33e3-4145-9fd5-026219c83b0f service nova] Lock "febeca1c-3370-4c74-8bd9-efad313df1e2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.781326] env[63372]: DEBUG oslo_concurrency.lockutils [req-6a718858-44a6-4b1a-863e-a57fd2c4ea60 req-6894fd71-33e3-4145-9fd5-026219c83b0f service nova] Lock "febeca1c-3370-4c74-8bd9-efad313df1e2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.781326] env[63372]: DEBUG nova.compute.manager [req-6a718858-44a6-4b1a-863e-a57fd2c4ea60 req-6894fd71-33e3-4145-9fd5-026219c83b0f service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] No waiting events found dispatching network-vif-plugged-485f2c37-5aa1-470e-8b39-a986ebd7fd24 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 767.781326] env[63372]: WARNING nova.compute.manager [req-6a718858-44a6-4b1a-863e-a57fd2c4ea60 req-6894fd71-33e3-4145-9fd5-026219c83b0f service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Received unexpected event network-vif-plugged-485f2c37-5aa1-470e-8b39-a986ebd7fd24 for instance with vm_state building and task_state spawning. [ 767.782733] env[63372]: INFO nova.compute.resource_tracker [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating resource usage from migration e4ca8d01-990e-4e1b-8d4a-067812994b8f [ 767.786878] env[63372]: DEBUG nova.compute.manager [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 767.932582] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "c01a5d24-eb46-4a69-993e-753880ce8e85" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.932876] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.966722] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023985, 'name': RemoveSnapshot_Task} progress is 75%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.992567] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023984, 'name': RelocateVM_Task} progress is 34%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.319980] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32223a95-5ead-411e-a4cf-d9640f580869 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.329141] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff1fd52-3c12-4d19-9c7a-2be261fd062c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.370848] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1acc60a5-d4e3-489a-bb29-2fd02d5bfb72 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.380464] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be91d516-0918-4cd8-973e-b138a7e9e9f0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.399101] env[63372]: DEBUG nova.compute.provider_tree [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.461683] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023985, 'name': RemoveSnapshot_Task} progress is 75%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.493345] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023984, 'name': RelocateVM_Task} progress is 49%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.584826] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "357505d0-f306-4e11-8a62-e03cfab2b7c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.585659] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "357505d0-f306-4e11-8a62-e03cfab2b7c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.796824] env[63372]: DEBUG nova.compute.manager [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 768.824885] env[63372]: DEBUG nova.virt.hardware [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 768.825313] env[63372]: DEBUG nova.virt.hardware [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 768.825552] env[63372]: DEBUG nova.virt.hardware [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 768.825710] env[63372]: DEBUG nova.virt.hardware [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 768.825924] env[63372]: DEBUG nova.virt.hardware [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 768.826156] env[63372]: DEBUG nova.virt.hardware [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 768.826393] env[63372]: DEBUG nova.virt.hardware [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 768.826572] env[63372]: DEBUG nova.virt.hardware [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 768.826770] env[63372]: DEBUG nova.virt.hardware [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 768.827026] env[63372]: DEBUG nova.virt.hardware [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 768.827182] env[63372]: DEBUG nova.virt.hardware [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 768.828402] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc68b59e-362e-467b-9a0b-60dcfe097362 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.838095] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d56035-8c28-4421-bf38-f8abbb3fb83d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.902059] env[63372]: DEBUG nova.scheduler.client.report [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.960379] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023985, 'name': RemoveSnapshot_Task} progress is 75%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.994625] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023984, 'name': RelocateVM_Task} progress is 62%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.074521] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "757df632-4af2-4eb3-bc6d-8c0812ba850d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.074898] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "757df632-4af2-4eb3-bc6d-8c0812ba850d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.075212] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "757df632-4af2-4eb3-bc6d-8c0812ba850d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.075582] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "757df632-4af2-4eb3-bc6d-8c0812ba850d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.075582] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "757df632-4af2-4eb3-bc6d-8c0812ba850d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.082332] env[63372]: INFO nova.compute.manager [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Terminating instance [ 769.085057] env[63372]: DEBUG nova.compute.manager [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 769.085354] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 769.086321] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc960dd-39dc-41c1-bbb0-a39aaff9cd18 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.094991] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 769.095420] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eae857b7-e791-41d2-9dfd-a35f02773acd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.103103] env[63372]: DEBUG oslo_vmware.api [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 769.103103] env[63372]: value = "task-1023986" [ 769.103103] env[63372]: _type = "Task" [ 769.103103] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.112041] env[63372]: DEBUG oslo_vmware.api [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023986, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.382885] env[63372]: DEBUG nova.network.neutron [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Successfully updated port: f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 769.407798] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.648s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.408115] env[63372]: INFO nova.compute.manager [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Migrating [ 769.408421] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.408616] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquired lock "compute-rpcapi-router" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.409976] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.537s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.411804] env[63372]: INFO nova.compute.claims [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.415513] env[63372]: INFO nova.compute.rpcapi [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 769.416376] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Releasing lock "compute-rpcapi-router" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.463398] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023985, 'name': RemoveSnapshot_Task} progress is 75%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.494620] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023984, 'name': RelocateVM_Task} progress is 75%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.556921] env[63372]: DEBUG nova.compute.manager [req-5636102e-784a-4fd0-a1f4-8517f6c0eb51 req-a22d8caf-3a83-4225-ba3e-5f632be939e5 service nova] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Received event network-vif-plugged-f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.557456] env[63372]: DEBUG oslo_concurrency.lockutils [req-5636102e-784a-4fd0-a1f4-8517f6c0eb51 req-a22d8caf-3a83-4225-ba3e-5f632be939e5 service nova] Acquiring lock "4c9eb955-3bed-4b26-866f-b2a876b835dd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.558107] env[63372]: DEBUG oslo_concurrency.lockutils [req-5636102e-784a-4fd0-a1f4-8517f6c0eb51 req-a22d8caf-3a83-4225-ba3e-5f632be939e5 service nova] Lock "4c9eb955-3bed-4b26-866f-b2a876b835dd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.558107] env[63372]: DEBUG oslo_concurrency.lockutils [req-5636102e-784a-4fd0-a1f4-8517f6c0eb51 req-a22d8caf-3a83-4225-ba3e-5f632be939e5 service nova] Lock "4c9eb955-3bed-4b26-866f-b2a876b835dd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.558107] env[63372]: DEBUG nova.compute.manager [req-5636102e-784a-4fd0-a1f4-8517f6c0eb51 req-a22d8caf-3a83-4225-ba3e-5f632be939e5 service nova] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] No waiting events found dispatching network-vif-plugged-f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 769.558411] env[63372]: WARNING nova.compute.manager [req-5636102e-784a-4fd0-a1f4-8517f6c0eb51 req-a22d8caf-3a83-4225-ba3e-5f632be939e5 service nova] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Received unexpected event network-vif-plugged-f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff for instance with vm_state building and task_state spawning. [ 769.614131] env[63372]: DEBUG oslo_vmware.api [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023986, 'name': PowerOffVM_Task, 'duration_secs': 0.270225} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.614586] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 769.614835] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 769.615258] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8fdb357f-861d-459d-852d-aa9cc03cc444 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.626662] env[63372]: DEBUG oslo_vmware.rw_handles [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5205f68b-4c16-0d99-acd7-8b608ec1b4e1/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 769.627730] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0bb9269-8094-46c4-8c08-f60125ec18a9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.635811] env[63372]: DEBUG oslo_vmware.rw_handles [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5205f68b-4c16-0d99-acd7-8b608ec1b4e1/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 769.636154] env[63372]: ERROR oslo_vmware.rw_handles [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5205f68b-4c16-0d99-acd7-8b608ec1b4e1/disk-0.vmdk due to incomplete transfer. [ 769.636371] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ae84749c-0aff-4b72-bcfc-74f299d598a5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.645500] env[63372]: DEBUG oslo_vmware.rw_handles [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5205f68b-4c16-0d99-acd7-8b608ec1b4e1/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 769.645818] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Uploaded image 3c9be794-7dcb-43df-a9b2-d9289ac8889f to the Glance image server {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 769.648823] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 769.649140] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6f46cb14-8a19-4dd7-8c24-f991363df496 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.657475] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 769.657475] env[63372]: value = "task-1023988" [ 769.657475] env[63372]: _type = "Task" [ 769.657475] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.666930] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023988, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.722982] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 769.723304] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 769.723495] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleting the datastore file [datastore1] 757df632-4af2-4eb3-bc6d-8c0812ba850d {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 769.724030] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-57f9b96d-9ad8-4616-9a61-7e29e532404d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.732863] env[63372]: DEBUG nova.compute.manager [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Received event network-changed-29995495-a43f-4be1-b907-c88d9fe78dcc {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.734081] env[63372]: DEBUG nova.compute.manager [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Refreshing instance network info cache due to event network-changed-29995495-a43f-4be1-b907-c88d9fe78dcc. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 769.734081] env[63372]: DEBUG oslo_concurrency.lockutils [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] Acquiring lock "refresh_cache-69a107d0-80c7-42e5-b514-b4273e1a3359" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.734313] env[63372]: DEBUG oslo_concurrency.lockutils [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] Acquired lock "refresh_cache-69a107d0-80c7-42e5-b514-b4273e1a3359" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.734377] env[63372]: DEBUG nova.network.neutron [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Refreshing network info cache for port 29995495-a43f-4be1-b907-c88d9fe78dcc {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 769.739221] env[63372]: DEBUG oslo_vmware.api [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 769.739221] env[63372]: value = "task-1023989" [ 769.739221] env[63372]: _type = "Task" [ 769.739221] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.751764] env[63372]: DEBUG oslo_vmware.api [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023989, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.889595] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Acquiring lock "refresh_cache-4c9eb955-3bed-4b26-866f-b2a876b835dd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.889595] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Acquired lock "refresh_cache-4c9eb955-3bed-4b26-866f-b2a876b835dd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.889595] env[63372]: DEBUG nova.network.neutron [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 769.935942] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.935942] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquired lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.935942] env[63372]: DEBUG nova.network.neutron [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 769.963232] env[63372]: DEBUG oslo_vmware.api [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1023985, 'name': RemoveSnapshot_Task, 'duration_secs': 2.325097} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.963301] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 769.963573] env[63372]: INFO nova.compute.manager [None req-8b515893-4f6f-4bb4-98f0-0a7b6d9622c2 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Took 18.75 seconds to snapshot the instance on the hypervisor. [ 769.996867] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023984, 'name': RelocateVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.168540] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023988, 'name': Destroy_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.251800] env[63372]: DEBUG oslo_vmware.api [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023989, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.410901] env[63372]: DEBUG nova.network.neutron [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Successfully updated port: 9249f3d1-43e4-4e00-94a1-d4ed0b20b171 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 770.446071] env[63372]: DEBUG nova.network.neutron [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.503559] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023984, 'name': RelocateVM_Task} progress is 97%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.530235] env[63372]: DEBUG nova.compute.manager [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 770.534019] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-848cfe04-6a6b-4f50-a2d3-c5aa6b2b1f43 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.551718] env[63372]: DEBUG nova.network.neutron [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Updated VIF entry in instance network info cache for port 29995495-a43f-4be1-b907-c88d9fe78dcc. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 770.551718] env[63372]: DEBUG nova.network.neutron [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Updating instance_info_cache with network_info: [{"id": "29995495-a43f-4be1-b907-c88d9fe78dcc", "address": "fa:16:3e:cb:ee:42", "network": {"id": "3ca584f0-a422-4ca2-bdc9-033462b8580e", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-175791853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "541cfc7f0fba40678e1f67d677d22df2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c6a4836-66dc-4e43-982b-f8fcd3f9989a", "external-id": "nsx-vlan-transportzone-635", "segmentation_id": 635, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29995495-a4", "ovs_interfaceid": "29995495-a43f-4be1-b907-c88d9fe78dcc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.633539] env[63372]: DEBUG nova.network.neutron [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Updating instance_info_cache with network_info: [{"id": "f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff", "address": "fa:16:3e:92:e5:5d", "network": {"id": "8167e8ec-3d8f-4a74-b1d7-32f81c078b63", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-859761753-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d9532ba6f40b4592bd29d2f78039377a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee", "external-id": "nsx-vlan-transportzone-401", "segmentation_id": 401, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7d8a5f6-6c", "ovs_interfaceid": "f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.671224] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023988, 'name': Destroy_Task, 'duration_secs': 0.616015} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.673784] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Destroyed the VM [ 770.674137] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 770.674581] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-fb324e1b-a52d-496d-9458-cebbd8f2bf4b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.681689] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 770.681689] env[63372]: value = "task-1023990" [ 770.681689] env[63372]: _type = "Task" [ 770.681689] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.704714] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023990, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.760410] env[63372]: DEBUG oslo_vmware.api [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1023989, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.55068} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.760801] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 770.761283] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 770.761404] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 770.761792] env[63372]: INFO nova.compute.manager [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Took 1.68 seconds to destroy the instance on the hypervisor. [ 770.761885] env[63372]: DEBUG oslo.service.loopingcall [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 770.762080] env[63372]: DEBUG nova.compute.manager [-] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.762200] env[63372]: DEBUG nova.network.neutron [-] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 770.769153] env[63372]: DEBUG nova.network.neutron [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance_info_cache with network_info: [{"id": "874f14b1-6cc3-4701-8a08-35940fc9bd9b", "address": "fa:16:3e:57:46:8e", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.24", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap874f14b1-6c", "ovs_interfaceid": "874f14b1-6cc3-4701-8a08-35940fc9bd9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.913921] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "refresh_cache-febeca1c-3370-4c74-8bd9-efad313df1e2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.914086] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquired lock "refresh_cache-febeca1c-3370-4c74-8bd9-efad313df1e2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.914241] env[63372]: DEBUG nova.network.neutron [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 770.994922] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023984, 'name': RelocateVM_Task} progress is 98%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.001703] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6f96cc-d0d0-4c97-b698-17352b73d8f7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.012581] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5a82f0-a25c-4e7e-a573-c5791e799e54 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.041821] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f4db54-679c-4ac7-8ee0-9affd57c9ba4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.051144] env[63372]: INFO nova.compute.manager [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] instance snapshotting [ 771.054861] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1465cd13-6e89-45de-847e-158f19d4e689 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.061412] env[63372]: DEBUG oslo_concurrency.lockutils [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] Releasing lock "refresh_cache-69a107d0-80c7-42e5-b514-b4273e1a3359" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.063538] env[63372]: DEBUG nova.compute.manager [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Received event network-changed-485f2c37-5aa1-470e-8b39-a986ebd7fd24 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.064992] env[63372]: DEBUG nova.compute.manager [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Refreshing instance network info cache due to event network-changed-485f2c37-5aa1-470e-8b39-a986ebd7fd24. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 771.064992] env[63372]: DEBUG oslo_concurrency.lockutils [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] Acquiring lock "refresh_cache-febeca1c-3370-4c74-8bd9-efad313df1e2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.064992] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eccfc90-bcaf-4120-ad39-9e61e2af5958 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.081911] env[63372]: DEBUG nova.compute.provider_tree [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.099309] env[63372]: DEBUG nova.scheduler.client.report [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.103943] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed4c554-65d6-44e4-93b4-0a5cb2d96db5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.136614] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Releasing lock "refresh_cache-4c9eb955-3bed-4b26-866f-b2a876b835dd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.136959] env[63372]: DEBUG nova.compute.manager [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Instance network_info: |[{"id": "f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff", "address": "fa:16:3e:92:e5:5d", "network": {"id": "8167e8ec-3d8f-4a74-b1d7-32f81c078b63", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-859761753-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d9532ba6f40b4592bd29d2f78039377a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee", "external-id": "nsx-vlan-transportzone-401", "segmentation_id": 401, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7d8a5f6-6c", "ovs_interfaceid": "f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 771.137600] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:e5:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 771.145960] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Creating folder: Project (d9532ba6f40b4592bd29d2f78039377a). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 771.146904] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd3ec0e2-f284-4837-a9b9-c00f0cd5cb68 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.158245] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Created folder: Project (d9532ba6f40b4592bd29d2f78039377a) in parent group-v227230. [ 771.158245] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Creating folder: Instances. Parent ref: group-v227336. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 771.158245] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-124d75a2-4694-4c72-b022-9258a238c350 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.167935] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Created folder: Instances in parent group-v227336. [ 771.168229] env[63372]: DEBUG oslo.service.loopingcall [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.168451] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 771.168721] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-941203b8-3bf7-41f6-8451-cb645091cd27 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.195573] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 771.195573] env[63372]: value = "task-1023993" [ 771.195573] env[63372]: _type = "Task" [ 771.195573] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.195821] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023990, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.199685] env[63372]: DEBUG oslo_concurrency.lockutils [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "7cf6ac9b-4703-4143-a22c-abc5528ce5ef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.199795] env[63372]: DEBUG oslo_concurrency.lockutils [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "7cf6ac9b-4703-4143-a22c-abc5528ce5ef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.199996] env[63372]: DEBUG oslo_concurrency.lockutils [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "7cf6ac9b-4703-4143-a22c-abc5528ce5ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.200241] env[63372]: DEBUG oslo_concurrency.lockutils [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "7cf6ac9b-4703-4143-a22c-abc5528ce5ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.200425] env[63372]: DEBUG oslo_concurrency.lockutils [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "7cf6ac9b-4703-4143-a22c-abc5528ce5ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.203065] env[63372]: INFO nova.compute.manager [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Terminating instance [ 771.208314] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023993, 'name': CreateVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.209021] env[63372]: DEBUG nova.compute.manager [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 771.209444] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 771.210284] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7458396-ac59-4459-94dd-a35385de2f11 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.218108] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 771.218639] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b798701b-e89a-47e7-9277-2403c812e9e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.270328] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Releasing lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.301659] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 771.302024] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 771.302140] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Deleting the datastore file [datastore1] 7cf6ac9b-4703-4143-a22c-abc5528ce5ef {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 771.302521] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c1f0d6f4-6edc-45b2-bf44-c6705422914f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.311276] env[63372]: DEBUG oslo_vmware.api [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 771.311276] env[63372]: value = "task-1023995" [ 771.311276] env[63372]: _type = "Task" [ 771.311276] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.318577] env[63372]: DEBUG oslo_vmware.api [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023995, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.453648] env[63372]: DEBUG nova.network.neutron [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 771.495609] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023984, 'name': RelocateVM_Task, 'duration_secs': 4.646167} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.495918] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Volume attach. Driver type: vmdk {{(pid=63372) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 771.496160] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227276', 'volume_id': '3226fd91-2732-41ce-bb8e-f909a830c8d4', 'name': 'volume-3226fd91-2732-41ce-bb8e-f909a830c8d4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '69a107d0-80c7-42e5-b514-b4273e1a3359', 'attached_at': '', 'detached_at': '', 'volume_id': '3226fd91-2732-41ce-bb8e-f909a830c8d4', 'serial': '3226fd91-2732-41ce-bb8e-f909a830c8d4'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 771.496940] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e61e87a-c3f2-4ab6-a36a-75ba12f73764 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.515618] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79a08a8e-57ff-4b7b-a2c1-a68ce0c2c2f5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.539637] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] volume-3226fd91-2732-41ce-bb8e-f909a830c8d4/volume-3226fd91-2732-41ce-bb8e-f909a830c8d4.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 771.542721] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5ec6e66-361e-474c-8510-7b3224564dd0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.556990] env[63372]: DEBUG nova.network.neutron [-] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.566223] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Waiting for the task: (returnval){ [ 771.566223] env[63372]: value = "task-1023996" [ 771.566223] env[63372]: _type = "Task" [ 771.566223] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.576891] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023996, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.606841] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.197s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.607998] env[63372]: DEBUG nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 771.610804] env[63372]: DEBUG oslo_concurrency.lockutils [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.017s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.610912] env[63372]: DEBUG oslo_concurrency.lockutils [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.612863] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.975s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.614998] env[63372]: INFO nova.compute.claims [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.618838] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 771.619348] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d380d659-31da-40c8-8180-d1b90732eff6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.626872] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 771.626872] env[63372]: value = "task-1023997" [ 771.626872] env[63372]: _type = "Task" [ 771.626872] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.637179] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023997, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.646619] env[63372]: INFO nova.scheduler.client.report [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Deleted allocations for instance 63b9095c-fb49-4dc1-a6dc-96529aeaab81 [ 771.697315] env[63372]: DEBUG oslo_vmware.api [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1023990, 'name': RemoveSnapshot_Task, 'duration_secs': 0.719475} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.700818] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 771.701127] env[63372]: INFO nova.compute.manager [None req-985909cc-d425-4fea-b53c-7b174562b170 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Took 16.85 seconds to snapshot the instance on the hypervisor. [ 771.712317] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1023993, 'name': CreateVM_Task, 'duration_secs': 0.3811} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.712317] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 771.713015] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.713255] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.713854] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 771.714151] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cc76ce1-b6c0-4aff-85b8-55a059419ce0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.721736] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Waiting for the task: (returnval){ [ 771.721736] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]526674dd-07f0-6b95-2a41-f07bb4c98d92" [ 771.721736] env[63372]: _type = "Task" [ 771.721736] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.736379] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526674dd-07f0-6b95-2a41-f07bb4c98d92, 'name': SearchDatastore_Task, 'duration_secs': 0.009099} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.737753] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.738248] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 771.738689] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.738947] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.739301] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 771.740074] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4fbd0968-5754-4d92-99ae-91c71e783b3f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.756525] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 771.756897] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 771.760096] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cce04c5d-5004-423d-ba72-b951963e409d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.766358] env[63372]: DEBUG nova.compute.manager [req-0f9f8e05-24c4-4202-ac69-b17e2da71a63 req-a8f588e6-cbf2-49ee-9b46-45ce14b8105d service nova] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Received event network-changed-f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.766510] env[63372]: DEBUG nova.compute.manager [req-0f9f8e05-24c4-4202-ac69-b17e2da71a63 req-a8f588e6-cbf2-49ee-9b46-45ce14b8105d service nova] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Refreshing instance network info cache due to event network-changed-f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 771.766868] env[63372]: DEBUG oslo_concurrency.lockutils [req-0f9f8e05-24c4-4202-ac69-b17e2da71a63 req-a8f588e6-cbf2-49ee-9b46-45ce14b8105d service nova] Acquiring lock "refresh_cache-4c9eb955-3bed-4b26-866f-b2a876b835dd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.767238] env[63372]: DEBUG oslo_concurrency.lockutils [req-0f9f8e05-24c4-4202-ac69-b17e2da71a63 req-a8f588e6-cbf2-49ee-9b46-45ce14b8105d service nova] Acquired lock "refresh_cache-4c9eb955-3bed-4b26-866f-b2a876b835dd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.769533] env[63372]: DEBUG nova.network.neutron [req-0f9f8e05-24c4-4202-ac69-b17e2da71a63 req-a8f588e6-cbf2-49ee-9b46-45ce14b8105d service nova] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Refreshing network info cache for port f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 771.783021] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Waiting for the task: (returnval){ [ 771.783021] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52c3f5b6-7a92-aec8-ea7e-b2967be81fd5" [ 771.783021] env[63372]: _type = "Task" [ 771.783021] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.796315] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c3f5b6-7a92-aec8-ea7e-b2967be81fd5, 'name': SearchDatastore_Task, 'duration_secs': 0.008269} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.803594] env[63372]: DEBUG nova.compute.manager [req-ea95a498-ede4-41fd-94bf-b1d069736d82 req-4131ceb1-9da7-4887-a82c-fe833dc9d528 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Received event network-vif-plugged-9249f3d1-43e4-4e00-94a1-d4ed0b20b171 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.803926] env[63372]: DEBUG oslo_concurrency.lockutils [req-ea95a498-ede4-41fd-94bf-b1d069736d82 req-4131ceb1-9da7-4887-a82c-fe833dc9d528 service nova] Acquiring lock "febeca1c-3370-4c74-8bd9-efad313df1e2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.804271] env[63372]: DEBUG oslo_concurrency.lockutils [req-ea95a498-ede4-41fd-94bf-b1d069736d82 req-4131ceb1-9da7-4887-a82c-fe833dc9d528 service nova] Lock "febeca1c-3370-4c74-8bd9-efad313df1e2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.804546] env[63372]: DEBUG oslo_concurrency.lockutils [req-ea95a498-ede4-41fd-94bf-b1d069736d82 req-4131ceb1-9da7-4887-a82c-fe833dc9d528 service nova] Lock "febeca1c-3370-4c74-8bd9-efad313df1e2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.804882] env[63372]: DEBUG nova.compute.manager [req-ea95a498-ede4-41fd-94bf-b1d069736d82 req-4131ceb1-9da7-4887-a82c-fe833dc9d528 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] No waiting events found dispatching network-vif-plugged-9249f3d1-43e4-4e00-94a1-d4ed0b20b171 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 771.805189] env[63372]: WARNING nova.compute.manager [req-ea95a498-ede4-41fd-94bf-b1d069736d82 req-4131ceb1-9da7-4887-a82c-fe833dc9d528 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Received unexpected event network-vif-plugged-9249f3d1-43e4-4e00-94a1-d4ed0b20b171 for instance with vm_state building and task_state spawning. [ 771.805452] env[63372]: DEBUG nova.compute.manager [req-ea95a498-ede4-41fd-94bf-b1d069736d82 req-4131ceb1-9da7-4887-a82c-fe833dc9d528 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Received event network-changed-9249f3d1-43e4-4e00-94a1-d4ed0b20b171 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.805719] env[63372]: DEBUG nova.compute.manager [req-ea95a498-ede4-41fd-94bf-b1d069736d82 req-4131ceb1-9da7-4887-a82c-fe833dc9d528 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Refreshing instance network info cache due to event network-changed-9249f3d1-43e4-4e00-94a1-d4ed0b20b171. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 771.806090] env[63372]: DEBUG oslo_concurrency.lockutils [req-ea95a498-ede4-41fd-94bf-b1d069736d82 req-4131ceb1-9da7-4887-a82c-fe833dc9d528 service nova] Acquiring lock "refresh_cache-febeca1c-3370-4c74-8bd9-efad313df1e2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.807130] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-097fa4b6-8240-44ea-926a-bf0b22a5efff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.823871] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Waiting for the task: (returnval){ [ 771.823871] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52c9ea9e-87e4-67ef-3dfe-626295121816" [ 771.823871] env[63372]: _type = "Task" [ 771.823871] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.829337] env[63372]: DEBUG oslo_vmware.api [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1023995, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.222631} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.834098] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 771.834649] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 771.834649] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 771.834877] env[63372]: INFO nova.compute.manager [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Took 0.63 seconds to destroy the instance on the hypervisor. [ 771.835177] env[63372]: DEBUG oslo.service.loopingcall [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.835820] env[63372]: DEBUG nova.compute.manager [-] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.836018] env[63372]: DEBUG nova.network.neutron [-] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 771.846626] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c9ea9e-87e4-67ef-3dfe-626295121816, 'name': SearchDatastore_Task, 'duration_secs': 0.009794} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.847014] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.847434] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 4c9eb955-3bed-4b26-866f-b2a876b835dd/4c9eb955-3bed-4b26-866f-b2a876b835dd.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 771.848690] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce34b456-fa03-46df-b12c-8235122bcb1f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.855341] env[63372]: DEBUG nova.network.neutron [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Updating instance_info_cache with network_info: [{"id": "485f2c37-5aa1-470e-8b39-a986ebd7fd24", "address": "fa:16:3e:ed:e6:26", "network": {"id": "d9a3bf76-fc50-415b-a6fd-d68bf2092a77", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1360681471", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap485f2c37-5a", "ovs_interfaceid": "485f2c37-5aa1-470e-8b39-a986ebd7fd24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9249f3d1-43e4-4e00-94a1-d4ed0b20b171", "address": "fa:16:3e:20:73:bd", "network": {"id": "5c52a962-a156-4955-9c26-9c98e8273cb5", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-89594243", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.167", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "67921bdb-a7a0-46b5-ba05-ca997496e222", "external-id": "nsx-vlan-transportzone-856", "segmentation_id": 856, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9249f3d1-43", "ovs_interfaceid": "9249f3d1-43e4-4e00-94a1-d4ed0b20b171", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.859329] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Waiting for the task: (returnval){ [ 771.859329] env[63372]: value = "task-1023998" [ 771.859329] env[63372]: _type = "Task" [ 771.859329] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.871151] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1023998, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.892387] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "368101ee-e4fd-4fe4-b7c8-75c6b210189f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.892795] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "368101ee-e4fd-4fe4-b7c8-75c6b210189f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.061610] env[63372]: INFO nova.compute.manager [-] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Took 1.30 seconds to deallocate network for instance. [ 772.076124] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023996, 'name': ReconfigVM_Task, 'duration_secs': 0.378738} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.076369] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Reconfigured VM instance instance-00000037 to attach disk [datastore2] volume-3226fd91-2732-41ce-bb8e-f909a830c8d4/volume-3226fd91-2732-41ce-bb8e-f909a830c8d4.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 772.081888] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-581f8464-0265-490a-837e-f58cb9394410 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.099861] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Waiting for the task: (returnval){ [ 772.099861] env[63372]: value = "task-1023999" [ 772.099861] env[63372]: _type = "Task" [ 772.099861] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.111378] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023999, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.119958] env[63372]: DEBUG nova.compute.utils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 772.124835] env[63372]: DEBUG nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 772.124888] env[63372]: DEBUG nova.network.neutron [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 772.141201] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023997, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.158435] env[63372]: DEBUG oslo_concurrency.lockutils [None req-917cd2b9-0de6-4523-8d34-30bcec50e074 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "63b9095c-fb49-4dc1-a6dc-96529aeaab81" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.581s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.188941] env[63372]: DEBUG nova.policy [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3d8052fc4294245bec22af4c254fce0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41ca5f8268434d329deec67fb2867b59', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 772.360492] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Releasing lock "refresh_cache-febeca1c-3370-4c74-8bd9-efad313df1e2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.361064] env[63372]: DEBUG nova.compute.manager [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Instance network_info: |[{"id": "485f2c37-5aa1-470e-8b39-a986ebd7fd24", "address": "fa:16:3e:ed:e6:26", "network": {"id": "d9a3bf76-fc50-415b-a6fd-d68bf2092a77", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1360681471", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap485f2c37-5a", "ovs_interfaceid": "485f2c37-5aa1-470e-8b39-a986ebd7fd24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9249f3d1-43e4-4e00-94a1-d4ed0b20b171", "address": "fa:16:3e:20:73:bd", "network": {"id": "5c52a962-a156-4955-9c26-9c98e8273cb5", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-89594243", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.167", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "67921bdb-a7a0-46b5-ba05-ca997496e222", "external-id": "nsx-vlan-transportzone-856", "segmentation_id": 856, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9249f3d1-43", "ovs_interfaceid": "9249f3d1-43e4-4e00-94a1-d4ed0b20b171", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 772.361607] env[63372]: DEBUG oslo_concurrency.lockutils [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] Acquired lock "refresh_cache-febeca1c-3370-4c74-8bd9-efad313df1e2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.361827] env[63372]: DEBUG nova.network.neutron [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Refreshing network info cache for port 485f2c37-5aa1-470e-8b39-a986ebd7fd24 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 772.363306] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ed:e6:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac7039c0-3374-4c08-87fc-af2449b48b02', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '485f2c37-5aa1-470e-8b39-a986ebd7fd24', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:73:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '67921bdb-a7a0-46b5-ba05-ca997496e222', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9249f3d1-43e4-4e00-94a1-d4ed0b20b171', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 772.374168] env[63372]: DEBUG oslo.service.loopingcall [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.381631] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 772.382184] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-119719e9-085d-4ec6-8db5-c2ad9b97ac41 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.408763] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1023998, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508986} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.414749] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 4c9eb955-3bed-4b26-866f-b2a876b835dd/4c9eb955-3bed-4b26-866f-b2a876b835dd.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 772.414749] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 772.414749] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 772.414749] env[63372]: value = "task-1024000" [ 772.414749] env[63372]: _type = "Task" [ 772.414749] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.414749] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3056471b-df61-47d0-95e9-dd93f092a14f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.426220] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024000, 'name': CreateVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.430027] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Waiting for the task: (returnval){ [ 772.430027] env[63372]: value = "task-1024001" [ 772.430027] env[63372]: _type = "Task" [ 772.430027] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.436329] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1024001, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.572511] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.612104] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023999, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.624801] env[63372]: DEBUG nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 772.640257] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1023997, 'name': CreateSnapshot_Task, 'duration_secs': 0.650647} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.640616] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 772.641431] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18434169-0b5a-4c3e-8580-f3b5e682e3f1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.714767] env[63372]: DEBUG nova.network.neutron [req-0f9f8e05-24c4-4202-ac69-b17e2da71a63 req-a8f588e6-cbf2-49ee-9b46-45ce14b8105d service nova] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Updated VIF entry in instance network info cache for port f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 772.714767] env[63372]: DEBUG nova.network.neutron [req-0f9f8e05-24c4-4202-ac69-b17e2da71a63 req-a8f588e6-cbf2-49ee-9b46-45ce14b8105d service nova] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Updating instance_info_cache with network_info: [{"id": "f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff", "address": "fa:16:3e:92:e5:5d", "network": {"id": "8167e8ec-3d8f-4a74-b1d7-32f81c078b63", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-859761753-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d9532ba6f40b4592bd29d2f78039377a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee", "external-id": "nsx-vlan-transportzone-401", "segmentation_id": 401, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7d8a5f6-6c", "ovs_interfaceid": "f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.729679] env[63372]: DEBUG nova.compute.manager [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 772.731093] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6701d325-f42b-4e3e-88f0-87c83cb2cba8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.736472] env[63372]: DEBUG nova.network.neutron [-] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.792807] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d90d5d-4461-487b-8125-10a6fca51773 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.816414] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance '8283b736-ad02-4082-97b7-561bd5c5da93' progress to 0 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 772.899801] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 772.900469] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 772.928567] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024000, 'name': CreateVM_Task, 'duration_secs': 0.404613} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.931451] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 772.934234] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.934414] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.934729] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 772.935666] env[63372]: DEBUG nova.network.neutron [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Successfully created port: 49b9f908-9366-4318-b2d3-c408e404a9df {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 772.939731] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8154ed67-57ef-456c-ac8e-47b310e76b94 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.948556] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1024001, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107267} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.949207] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 772.953404] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d99057c-344b-4273-8a80-e07619fa5900 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.958542] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 772.958542] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f84130-990b-0ef9-1ef2-3c03235a1151" [ 772.958542] env[63372]: _type = "Task" [ 772.958542] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.983200] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 4c9eb955-3bed-4b26-866f-b2a876b835dd/4c9eb955-3bed-4b26-866f-b2a876b835dd.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 772.986666] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9a6e95a-38cc-4f2a-8e05-bae31925d87c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.006122] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f84130-990b-0ef9-1ef2-3c03235a1151, 'name': SearchDatastore_Task, 'duration_secs': 0.011428} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.006932] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.007268] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 773.007562] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.007763] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.007968] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 773.008266] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b1cd21ae-ee2d-419a-98c6-618045196434 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.011544] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Waiting for the task: (returnval){ [ 773.011544] env[63372]: value = "task-1024002" [ 773.011544] env[63372]: _type = "Task" [ 773.011544] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.012420] env[63372]: DEBUG nova.network.neutron [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Updated VIF entry in instance network info cache for port 485f2c37-5aa1-470e-8b39-a986ebd7fd24. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 773.012840] env[63372]: DEBUG nova.network.neutron [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Updating instance_info_cache with network_info: [{"id": "485f2c37-5aa1-470e-8b39-a986ebd7fd24", "address": "fa:16:3e:ed:e6:26", "network": {"id": "d9a3bf76-fc50-415b-a6fd-d68bf2092a77", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1360681471", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap485f2c37-5a", "ovs_interfaceid": "485f2c37-5aa1-470e-8b39-a986ebd7fd24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9249f3d1-43e4-4e00-94a1-d4ed0b20b171", "address": "fa:16:3e:20:73:bd", "network": {"id": "5c52a962-a156-4955-9c26-9c98e8273cb5", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-89594243", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.167", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "67921bdb-a7a0-46b5-ba05-ca997496e222", "external-id": "nsx-vlan-transportzone-856", "segmentation_id": 856, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9249f3d1-43", "ovs_interfaceid": "9249f3d1-43e4-4e00-94a1-d4ed0b20b171", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.023351] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 773.023351] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 773.023351] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21eb71dd-0a14-4384-ad84-2ffe0d2c15da {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.029461] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1024002, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.033511] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 773.033511] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]522e8f24-b837-4310-bc77-16687426b746" [ 773.033511] env[63372]: _type = "Task" [ 773.033511] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.045315] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522e8f24-b837-4310-bc77-16687426b746, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.111481] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023999, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.172268] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 773.173750] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8288e24f-e89a-44ad-aa87-db0f9b7bdad6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.184538] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 773.184538] env[63372]: value = "task-1024003" [ 773.184538] env[63372]: _type = "Task" [ 773.184538] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.201329] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024003, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.218022] env[63372]: DEBUG oslo_concurrency.lockutils [req-0f9f8e05-24c4-4202-ac69-b17e2da71a63 req-a8f588e6-cbf2-49ee-9b46-45ce14b8105d service nova] Releasing lock "refresh_cache-4c9eb955-3bed-4b26-866f-b2a876b835dd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.218342] env[63372]: DEBUG nova.compute.manager [req-0f9f8e05-24c4-4202-ac69-b17e2da71a63 req-a8f588e6-cbf2-49ee-9b46-45ce14b8105d service nova] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Received event network-vif-deleted-f6d823dd-d794-4292-8906-9988287ccff0 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.241770] env[63372]: INFO nova.compute.manager [-] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Took 1.41 seconds to deallocate network for instance. [ 773.242455] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d175dc0-008c-4e22-b2c3-c59392fa89b5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.256995] env[63372]: INFO nova.compute.manager [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] instance snapshotting [ 773.262778] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b7ca33-b158-40cb-be2a-17df128655ea {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.266988] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e7d393a-f36d-4b62-a04b-56310ab5feca {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.318338] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e5ddf91-1cbe-4eed-8e0d-393935f4ae6a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.321555] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c987e0-f499-49f9-b261-c8ef4c2ba394 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.327160] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 773.327160] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-94125bfc-a1e2-4481-b5ed-b2d711d969a2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.337539] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bccef83a-0663-4f43-9d68-5a2d3e7fb08b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.342619] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 773.342619] env[63372]: value = "task-1024004" [ 773.342619] env[63372]: _type = "Task" [ 773.342619] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.354496] env[63372]: DEBUG nova.compute.provider_tree [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.360947] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024004, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.414362] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 773.414362] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Starting heal instance info cache {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 773.523845] env[63372]: DEBUG oslo_concurrency.lockutils [req-405c0dac-c2a1-4061-b339-9b798c6b059e req-46513de3-57ac-4a1c-a22a-1f3c6bf82ec4 service nova] Releasing lock "refresh_cache-febeca1c-3370-4c74-8bd9-efad313df1e2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.524518] env[63372]: DEBUG oslo_concurrency.lockutils [req-ea95a498-ede4-41fd-94bf-b1d069736d82 req-4131ceb1-9da7-4887-a82c-fe833dc9d528 service nova] Acquired lock "refresh_cache-febeca1c-3370-4c74-8bd9-efad313df1e2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.524779] env[63372]: DEBUG nova.network.neutron [req-ea95a498-ede4-41fd-94bf-b1d069736d82 req-4131ceb1-9da7-4887-a82c-fe833dc9d528 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Refreshing network info cache for port 9249f3d1-43e4-4e00-94a1-d4ed0b20b171 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 773.532208] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1024002, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.544251] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522e8f24-b837-4310-bc77-16687426b746, 'name': SearchDatastore_Task, 'duration_secs': 0.043524} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.545204] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e464792b-376f-46e9-b40e-ec28746233ab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.551561] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 773.551561] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521b69e8-9fb5-bf8f-7bd7-408779404332" [ 773.551561] env[63372]: _type = "Task" [ 773.551561] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.560234] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521b69e8-9fb5-bf8f-7bd7-408779404332, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.626025] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1023999, 'name': ReconfigVM_Task, 'duration_secs': 1.318098} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.626428] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227276', 'volume_id': '3226fd91-2732-41ce-bb8e-f909a830c8d4', 'name': 'volume-3226fd91-2732-41ce-bb8e-f909a830c8d4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '69a107d0-80c7-42e5-b514-b4273e1a3359', 'attached_at': '', 'detached_at': '', 'volume_id': '3226fd91-2732-41ce-bb8e-f909a830c8d4', 'serial': '3226fd91-2732-41ce-bb8e-f909a830c8d4'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 773.628129] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de2dace4-076b-4768-8670-12ce3b1997e2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.638501] env[63372]: DEBUG nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 773.649360] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Waiting for the task: (returnval){ [ 773.649360] env[63372]: value = "task-1024005" [ 773.649360] env[63372]: _type = "Task" [ 773.649360] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.672881] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1024005, 'name': Rename_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.705538] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 773.705836] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 773.706021] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 773.706267] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 773.706439] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 773.706610] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 773.706875] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 773.707079] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 773.707287] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 773.708872] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 773.709066] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 773.710082] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db7e762-1063-424c-a86d-dea825b1e6de {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.735346] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024003, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.735815] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01426b3d-a922-404e-b9af-5e5e888e50e7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.764034] env[63372]: DEBUG oslo_concurrency.lockutils [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.829264] env[63372]: DEBUG nova.compute.manager [req-b292a2dd-56e0-4167-9975-1182c9dc8de5 req-d9c72eb5-c439-4b74-9ca5-d248a7a7d337 service nova] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Received event network-vif-deleted-efd54f40-2ca1-4d06-a8c7-428f7508abe9 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.839231] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 773.839231] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-59e90bdf-9268-42f0-b46f-1e6aab29257f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.848590] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 773.848590] env[63372]: value = "task-1024006" [ 773.848590] env[63372]: _type = "Task" [ 773.848590] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.856362] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024004, 'name': PowerOffVM_Task, 'duration_secs': 0.241208} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.857923] env[63372]: DEBUG nova.scheduler.client.report [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.861625] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 773.862027] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance '8283b736-ad02-4082-97b7-561bd5c5da93' progress to 17 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 773.869433] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024006, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.948391] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.948538] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquired lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.948735] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Forcefully refreshing network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 774.026971] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1024002, 'name': ReconfigVM_Task, 'duration_secs': 0.541575} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.030621] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 4c9eb955-3bed-4b26-866f-b2a876b835dd/4c9eb955-3bed-4b26-866f-b2a876b835dd.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 774.030962] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-110fa744-ca25-42f9-81dc-8471ab32c9c5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.039487] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Waiting for the task: (returnval){ [ 774.039487] env[63372]: value = "task-1024007" [ 774.039487] env[63372]: _type = "Task" [ 774.039487] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.047812] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1024007, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.061557] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521b69e8-9fb5-bf8f-7bd7-408779404332, 'name': SearchDatastore_Task, 'duration_secs': 0.011963} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.061916] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.062256] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] febeca1c-3370-4c74-8bd9-efad313df1e2/febeca1c-3370-4c74-8bd9-efad313df1e2.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 774.062568] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-91b20f10-7018-4ee0-8c0b-4782e206c6b1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.072104] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 774.072104] env[63372]: value = "task-1024008" [ 774.072104] env[63372]: _type = "Task" [ 774.072104] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.079808] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024008, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.160813] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1024005, 'name': Rename_Task, 'duration_secs': 0.146743} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.162329] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 774.162329] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c33a6fd-02c3-45d3-8679-6e49c81c4245 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.168405] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Waiting for the task: (returnval){ [ 774.168405] env[63372]: value = "task-1024009" [ 774.168405] env[63372]: _type = "Task" [ 774.168405] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.178587] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1024009, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.203599] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024003, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.296284] env[63372]: DEBUG nova.network.neutron [req-ea95a498-ede4-41fd-94bf-b1d069736d82 req-4131ceb1-9da7-4887-a82c-fe833dc9d528 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Updated VIF entry in instance network info cache for port 9249f3d1-43e4-4e00-94a1-d4ed0b20b171. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 774.296800] env[63372]: DEBUG nova.network.neutron [req-ea95a498-ede4-41fd-94bf-b1d069736d82 req-4131ceb1-9da7-4887-a82c-fe833dc9d528 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Updating instance_info_cache with network_info: [{"id": "485f2c37-5aa1-470e-8b39-a986ebd7fd24", "address": "fa:16:3e:ed:e6:26", "network": {"id": "d9a3bf76-fc50-415b-a6fd-d68bf2092a77", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1360681471", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac7039c0-3374-4c08-87fc-af2449b48b02", "external-id": "nsx-vlan-transportzone-592", "segmentation_id": 592, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap485f2c37-5a", "ovs_interfaceid": "485f2c37-5aa1-470e-8b39-a986ebd7fd24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9249f3d1-43e4-4e00-94a1-d4ed0b20b171", "address": "fa:16:3e:20:73:bd", "network": {"id": "5c52a962-a156-4955-9c26-9c98e8273cb5", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-89594243", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.167", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "056172ef521c4de08b5033057773b2ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "67921bdb-a7a0-46b5-ba05-ca997496e222", "external-id": "nsx-vlan-transportzone-856", "segmentation_id": 856, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9249f3d1-43", "ovs_interfaceid": "9249f3d1-43e4-4e00-94a1-d4ed0b20b171", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.360638] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024006, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.366651] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.754s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.367162] env[63372]: DEBUG nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 774.370097] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.522s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.372048] env[63372]: INFO nova.compute.claims [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 774.376922] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 774.377250] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 774.378077] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 774.378077] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 774.378077] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 774.379344] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 774.379344] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 774.379344] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 774.379344] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 774.379576] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 774.383022] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 774.386743] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5dbcfcc1-934a-46c3-9bfe-75f483f95d56 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.407319] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 774.407319] env[63372]: value = "task-1024010" [ 774.407319] env[63372]: _type = "Task" [ 774.407319] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.417710] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024010, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.508793] env[63372]: DEBUG nova.compute.manager [req-d8680e22-dc79-4bed-ba18-90d90071b3ea req-b563260e-315e-4008-8991-237b7bca1be9 service nova] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Received event network-vif-plugged-49b9f908-9366-4318-b2d3-c408e404a9df {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 774.509360] env[63372]: DEBUG oslo_concurrency.lockutils [req-d8680e22-dc79-4bed-ba18-90d90071b3ea req-b563260e-315e-4008-8991-237b7bca1be9 service nova] Acquiring lock "bad91d77-c7ee-4572-b1ed-068b2a55233c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.509551] env[63372]: DEBUG oslo_concurrency.lockutils [req-d8680e22-dc79-4bed-ba18-90d90071b3ea req-b563260e-315e-4008-8991-237b7bca1be9 service nova] Lock "bad91d77-c7ee-4572-b1ed-068b2a55233c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.510019] env[63372]: DEBUG oslo_concurrency.lockutils [req-d8680e22-dc79-4bed-ba18-90d90071b3ea req-b563260e-315e-4008-8991-237b7bca1be9 service nova] Lock "bad91d77-c7ee-4572-b1ed-068b2a55233c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.510442] env[63372]: DEBUG nova.compute.manager [req-d8680e22-dc79-4bed-ba18-90d90071b3ea req-b563260e-315e-4008-8991-237b7bca1be9 service nova] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] No waiting events found dispatching network-vif-plugged-49b9f908-9366-4318-b2d3-c408e404a9df {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 774.510786] env[63372]: WARNING nova.compute.manager [req-d8680e22-dc79-4bed-ba18-90d90071b3ea req-b563260e-315e-4008-8991-237b7bca1be9 service nova] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Received unexpected event network-vif-plugged-49b9f908-9366-4318-b2d3-c408e404a9df for instance with vm_state building and task_state spawning. [ 774.549798] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1024007, 'name': Rename_Task, 'duration_secs': 0.307386} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.550128] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 774.550419] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ae7e029d-eaec-4bef-ae55-1aec6c97d0c3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.558419] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Waiting for the task: (returnval){ [ 774.558419] env[63372]: value = "task-1024011" [ 774.558419] env[63372]: _type = "Task" [ 774.558419] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.567412] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1024011, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.582140] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024008, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.626098] env[63372]: DEBUG nova.network.neutron [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Successfully updated port: 49b9f908-9366-4318-b2d3-c408e404a9df {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 774.680342] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1024009, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.703470] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024003, 'name': CloneVM_Task} progress is 95%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.801332] env[63372]: DEBUG oslo_concurrency.lockutils [req-ea95a498-ede4-41fd-94bf-b1d069736d82 req-4131ceb1-9da7-4887-a82c-fe833dc9d528 service nova] Releasing lock "refresh_cache-febeca1c-3370-4c74-8bd9-efad313df1e2" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.864956] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024006, 'name': CreateSnapshot_Task, 'duration_secs': 0.939002} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.865226] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 774.866194] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d52bc3-cd19-4525-a04b-c2b7b1570e3a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.877919] env[63372]: DEBUG nova.compute.utils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.881988] env[63372]: DEBUG nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 774.881988] env[63372]: DEBUG nova.network.neutron [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 774.918082] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024010, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.930075] env[63372]: DEBUG nova.policy [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3d8052fc4294245bec22af4c254fce0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41ca5f8268434d329deec67fb2867b59', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 775.011505] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "89301344-84a9-4d13-aae7-99943d0a478e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.011788] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "89301344-84a9-4d13-aae7-99943d0a478e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.068722] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1024011, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.081253] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024008, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.704817} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.083610] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] febeca1c-3370-4c74-8bd9-efad313df1e2/febeca1c-3370-4c74-8bd9-efad313df1e2.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 775.083836] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 775.084102] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-539562d9-0912-47f2-b295-02eaf2c7f7e0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.090535] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 775.090535] env[63372]: value = "task-1024012" [ 775.090535] env[63372]: _type = "Task" [ 775.090535] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.098587] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024012, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.129425] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "refresh_cache-bad91d77-c7ee-4572-b1ed-068b2a55233c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.129607] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired lock "refresh_cache-bad91d77-c7ee-4572-b1ed-068b2a55233c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.129803] env[63372]: DEBUG nova.network.neutron [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 775.179429] env[63372]: DEBUG oslo_vmware.api [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1024009, 'name': PowerOnVM_Task, 'duration_secs': 0.594008} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.182354] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 775.182354] env[63372]: INFO nova.compute.manager [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Took 9.69 seconds to spawn the instance on the hypervisor. [ 775.182354] env[63372]: DEBUG nova.compute.manager [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 775.183163] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827fcb6f-9089-450d-ac62-f712c872fe12 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.202693] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024003, 'name': CloneVM_Task, 'duration_secs': 2.014744} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.202896] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Created linked-clone VM from snapshot [ 775.204874] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155034f3-7333-4f22-b5b2-1af1072a94cd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.213842] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Uploading image 442ec676-c4b4-4d9a-80ab-4267edc97c35 {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 775.241624] env[63372]: DEBUG oslo_vmware.rw_handles [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 775.241624] env[63372]: value = "vm-227341" [ 775.241624] env[63372]: _type = "VirtualMachine" [ 775.241624] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 775.241899] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d54a14df-a0b3-4ba8-abe9-b690ceb3dd6b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.250326] env[63372]: DEBUG oslo_vmware.rw_handles [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Lease: (returnval){ [ 775.250326] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521e0498-1125-dabc-d3b2-8cea7a1759c6" [ 775.250326] env[63372]: _type = "HttpNfcLease" [ 775.250326] env[63372]: } obtained for exporting VM: (result){ [ 775.250326] env[63372]: value = "vm-227341" [ 775.250326] env[63372]: _type = "VirtualMachine" [ 775.250326] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 775.250590] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the lease: (returnval){ [ 775.250590] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521e0498-1125-dabc-d3b2-8cea7a1759c6" [ 775.250590] env[63372]: _type = "HttpNfcLease" [ 775.250590] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 775.257949] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 775.257949] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521e0498-1125-dabc-d3b2-8cea7a1759c6" [ 775.257949] env[63372]: _type = "HttpNfcLease" [ 775.257949] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 775.347329] env[63372]: DEBUG nova.network.neutron [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Successfully created port: 116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 775.353258] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Updating instance_info_cache with network_info: [{"id": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "address": "fa:16:3e:17:c8:2e", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4cf27a3-fc", "ovs_interfaceid": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.389719] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 775.390327] env[63372]: DEBUG nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 775.396946] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-45c9c7cc-0ae6-4895-9e89-e0b0ed1cdb90 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.406712] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 775.406712] env[63372]: value = "task-1024014" [ 775.406712] env[63372]: _type = "Task" [ 775.406712] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.422861] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024010, 'name': ReconfigVM_Task, 'duration_secs': 0.54547} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.427189] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance '8283b736-ad02-4082-97b7-561bd5c5da93' progress to 33 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 775.431133] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024014, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.573942] env[63372]: DEBUG oslo_vmware.api [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1024011, 'name': PowerOnVM_Task, 'duration_secs': 0.870439} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.577625] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 775.577625] env[63372]: INFO nova.compute.manager [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Took 6.78 seconds to spawn the instance on the hypervisor. [ 775.577625] env[63372]: DEBUG nova.compute.manager [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 775.578243] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d30f1b1-3ca8-4bbc-b025-557e5250047f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.605701] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024012, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.121225} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.606693] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 775.607585] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29ede87-e494-4bc7-847a-15e78eefca63 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.637116] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] febeca1c-3370-4c74-8bd9-efad313df1e2/febeca1c-3370-4c74-8bd9-efad313df1e2.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.646190] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62d6f4f1-8a0d-4662-9cac-0f2d9b356531 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.667328] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 775.667328] env[63372]: value = "task-1024015" [ 775.667328] env[63372]: _type = "Task" [ 775.667328] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.675235] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024015, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.693776] env[63372]: DEBUG nova.network.neutron [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 775.704040] env[63372]: INFO nova.compute.manager [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Took 46.37 seconds to build instance. [ 775.761171] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 775.761171] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521e0498-1125-dabc-d3b2-8cea7a1759c6" [ 775.761171] env[63372]: _type = "HttpNfcLease" [ 775.761171] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 775.761695] env[63372]: DEBUG oslo_vmware.rw_handles [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 775.761695] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521e0498-1125-dabc-d3b2-8cea7a1759c6" [ 775.761695] env[63372]: _type = "HttpNfcLease" [ 775.761695] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 775.762256] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6892cdc5-d406-495d-8d1b-770e660e68b4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.773104] env[63372]: DEBUG oslo_vmware.rw_handles [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526cdd31-0293-75ee-5fce-bdd2599d8960/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 775.773312] env[63372]: DEBUG oslo_vmware.rw_handles [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526cdd31-0293-75ee-5fce-bdd2599d8960/disk-0.vmdk for reading. {{(pid=63372) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 775.854934] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Releasing lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.855202] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Updated the network info_cache for instance {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 775.855418] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 775.855666] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 775.855868] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 775.856041] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 775.856237] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 775.856424] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 775.856666] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63372) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 775.856733] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager.update_available_resource {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 775.911763] env[63372]: DEBUG nova.network.neutron [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Updating instance_info_cache with network_info: [{"id": "49b9f908-9366-4318-b2d3-c408e404a9df", "address": "fa:16:3e:82:b4:d8", "network": {"id": "49776280-f4e5-4f22-98fd-a6e6ff70c80f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1268791360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41ca5f8268434d329deec67fb2867b59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap49b9f908-93", "ovs_interfaceid": "49b9f908-9366-4318-b2d3-c408e404a9df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.918746] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024014, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.925761] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a1ae647a-66d9-4cf8-b2da-50fdd02eddf6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.936981] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:36:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='51633402-95cc-485c-8808-65539d485326',id=27,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-443337402',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 775.937309] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 775.937471] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.937653] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 775.937828] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.937978] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 775.938214] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 775.938391] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 775.938571] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 775.938739] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 775.938912] env[63372]: DEBUG nova.virt.hardware [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 775.945534] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Reconfiguring VM instance instance-0000002f to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 775.950466] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5024bf51-4462-4361-84b8-f4f775e7da8b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.972047] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 775.972047] env[63372]: value = "task-1024016" [ 775.972047] env[63372]: _type = "Task" [ 775.972047] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.989361] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024016, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.035875] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad67b9ab-dd85-457e-afb9-1d5d8635a17a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.043878] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968f2b54-9683-41d7-9c54-443005156304 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.079257] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87551f1f-372e-40ed-9b2f-053f0651f488 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.086883] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4316c810-3d25-4ad4-99aa-b81e0306b523 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.104764] env[63372]: DEBUG nova.compute.provider_tree [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.107899] env[63372]: INFO nova.compute.manager [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Took 40.97 seconds to build instance. [ 776.177687] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024015, 'name': ReconfigVM_Task, 'duration_secs': 0.42871} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.177930] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Reconfigured VM instance instance-00000038 to attach disk [datastore2] febeca1c-3370-4c74-8bd9-efad313df1e2/febeca1c-3370-4c74-8bd9-efad313df1e2.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.178614] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-273e25da-ba0d-46aa-9d4c-8a8a900c057c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.185531] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 776.185531] env[63372]: value = "task-1024017" [ 776.185531] env[63372]: _type = "Task" [ 776.185531] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.194346] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024017, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.206162] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85daa173-0ae7-48df-b737-7d13af1f5c2a tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Lock "69a107d0-80c7-42e5-b514-b4273e1a3359" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.830s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.359961] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.407468] env[63372]: DEBUG nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 776.419369] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Releasing lock "refresh_cache-bad91d77-c7ee-4572-b1ed-068b2a55233c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.419547] env[63372]: DEBUG nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Instance network_info: |[{"id": "49b9f908-9366-4318-b2d3-c408e404a9df", "address": "fa:16:3e:82:b4:d8", "network": {"id": "49776280-f4e5-4f22-98fd-a6e6ff70c80f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1268791360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41ca5f8268434d329deec67fb2867b59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap49b9f908-93", "ovs_interfaceid": "49b9f908-9366-4318-b2d3-c408e404a9df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 776.419856] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024014, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.420238] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:b4:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffcecdaa-a7b8-49fc-9371-dbdb7744688e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '49b9f908-9366-4318-b2d3-c408e404a9df', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 776.428803] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Creating folder: Project (41ca5f8268434d329deec67fb2867b59). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 776.431360] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6bb264cb-381d-4170-86a4-731deaab27c6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.440331] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.440726] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.440962] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.441226] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.441472] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.441669] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.441913] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.442206] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.442456] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.443734] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.443734] env[63372]: DEBUG nova.virt.hardware [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.443916] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7743360c-cc96-4f18-8b00-54a936fba201 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.449901] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Created folder: Project (41ca5f8268434d329deec67fb2867b59) in parent group-v227230. [ 776.450402] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Creating folder: Instances. Parent ref: group-v227344. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 776.451144] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aa7f15a7-c9cb-4b50-bcb2-c04151e84f4a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.456987] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb5b705-bf42-4370-9228-81f97d205aa5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.463017] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Created folder: Instances in parent group-v227344. [ 776.463344] env[63372]: DEBUG oslo.service.loopingcall [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.473433] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 776.473971] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-83f29c1c-cd32-463f-90a1-f0f390a7dd6b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.498409] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024016, 'name': ReconfigVM_Task, 'duration_secs': 0.311522} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.500059] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Reconfigured VM instance instance-0000002f to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 776.500583] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 776.500583] env[63372]: value = "task-1024020" [ 776.500583] env[63372]: _type = "Task" [ 776.500583] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.501397] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6684cdd3-3e29-45f2-8229-6930f4a58a7a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.513024] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024020, 'name': CreateVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.532026] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] 8283b736-ad02-4082-97b7-561bd5c5da93/8283b736-ad02-4082-97b7-561bd5c5da93.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 776.532203] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90c83e68-9cb4-4c27-8214-8ef25efce514 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.548397] env[63372]: DEBUG nova.compute.manager [req-dadd4c5c-e765-49dc-9b87-bad654a508b3 req-5c61ac36-7604-49ea-b651-a6c0436ad763 service nova] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Received event network-changed-49b9f908-9366-4318-b2d3-c408e404a9df {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.548681] env[63372]: DEBUG nova.compute.manager [req-dadd4c5c-e765-49dc-9b87-bad654a508b3 req-5c61ac36-7604-49ea-b651-a6c0436ad763 service nova] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Refreshing instance network info cache due to event network-changed-49b9f908-9366-4318-b2d3-c408e404a9df. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 776.549299] env[63372]: DEBUG oslo_concurrency.lockutils [req-dadd4c5c-e765-49dc-9b87-bad654a508b3 req-5c61ac36-7604-49ea-b651-a6c0436ad763 service nova] Acquiring lock "refresh_cache-bad91d77-c7ee-4572-b1ed-068b2a55233c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.549299] env[63372]: DEBUG oslo_concurrency.lockutils [req-dadd4c5c-e765-49dc-9b87-bad654a508b3 req-5c61ac36-7604-49ea-b651-a6c0436ad763 service nova] Acquired lock "refresh_cache-bad91d77-c7ee-4572-b1ed-068b2a55233c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.549393] env[63372]: DEBUG nova.network.neutron [req-dadd4c5c-e765-49dc-9b87-bad654a508b3 req-5c61ac36-7604-49ea-b651-a6c0436ad763 service nova] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Refreshing network info cache for port 49b9f908-9366-4318-b2d3-c408e404a9df {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 776.555570] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 776.555570] env[63372]: value = "task-1024021" [ 776.555570] env[63372]: _type = "Task" [ 776.555570] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.565350] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024021, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.609682] env[63372]: DEBUG nova.scheduler.client.report [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.613164] env[63372]: DEBUG oslo_concurrency.lockutils [None req-215d512e-ee06-42bc-b666-27b75187932e tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Lock "4c9eb955-3bed-4b26-866f-b2a876b835dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.936s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.699604] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024017, 'name': Rename_Task, 'duration_secs': 0.173451} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.700021] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 776.700361] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13e233fe-01ea-4162-a15f-4fa6a0724288 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.708041] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 776.708041] env[63372]: value = "task-1024022" [ 776.708041] env[63372]: _type = "Task" [ 776.708041] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.708478] env[63372]: DEBUG nova.compute.manager [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 776.724266] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024022, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.872219] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "ac90a156-be00-4f62-a76e-e08914531167" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.872219] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "ac90a156-be00-4f62-a76e-e08914531167" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.922733] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024014, 'name': CloneVM_Task} progress is 95%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.003687] env[63372]: DEBUG nova.compute.manager [req-e9846c3f-ae32-4c6f-b893-74c98ebc773c req-49a1b8cf-8413-4f63-9f44-1873df48c8da service nova] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Received event network-vif-plugged-116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 777.003859] env[63372]: DEBUG oslo_concurrency.lockutils [req-e9846c3f-ae32-4c6f-b893-74c98ebc773c req-49a1b8cf-8413-4f63-9f44-1873df48c8da service nova] Acquiring lock "cbd55518-a3b2-4636-ba43-c279fdd0bf8d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.004096] env[63372]: DEBUG oslo_concurrency.lockutils [req-e9846c3f-ae32-4c6f-b893-74c98ebc773c req-49a1b8cf-8413-4f63-9f44-1873df48c8da service nova] Lock "cbd55518-a3b2-4636-ba43-c279fdd0bf8d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.004273] env[63372]: DEBUG oslo_concurrency.lockutils [req-e9846c3f-ae32-4c6f-b893-74c98ebc773c req-49a1b8cf-8413-4f63-9f44-1873df48c8da service nova] Lock "cbd55518-a3b2-4636-ba43-c279fdd0bf8d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.004436] env[63372]: DEBUG nova.compute.manager [req-e9846c3f-ae32-4c6f-b893-74c98ebc773c req-49a1b8cf-8413-4f63-9f44-1873df48c8da service nova] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] No waiting events found dispatching network-vif-plugged-116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 777.005017] env[63372]: WARNING nova.compute.manager [req-e9846c3f-ae32-4c6f-b893-74c98ebc773c req-49a1b8cf-8413-4f63-9f44-1873df48c8da service nova] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Received unexpected event network-vif-plugged-116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee for instance with vm_state building and task_state spawning. [ 777.022229] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024020, 'name': CreateVM_Task, 'duration_secs': 0.484241} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.022717] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 777.023866] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.026037] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.026037] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 777.026037] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a5c16ed-f6ab-4b12-87a5-e468109a0c1c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.031742] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 777.031742] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f2f691-49a2-b052-b8f6-5511ff184cca" [ 777.031742] env[63372]: _type = "Task" [ 777.031742] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.042240] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f2f691-49a2-b052-b8f6-5511ff184cca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.068074] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024021, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.115948] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.745s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.117800] env[63372]: DEBUG nova.compute.manager [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 777.122529] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.340s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.124799] env[63372]: INFO nova.compute.claims [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.128660] env[63372]: DEBUG nova.compute.manager [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 777.230673] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024022, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.248986] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.313592] env[63372]: DEBUG nova.network.neutron [req-dadd4c5c-e765-49dc-9b87-bad654a508b3 req-5c61ac36-7604-49ea-b651-a6c0436ad763 service nova] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Updated VIF entry in instance network info cache for port 49b9f908-9366-4318-b2d3-c408e404a9df. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 777.314477] env[63372]: DEBUG nova.network.neutron [req-dadd4c5c-e765-49dc-9b87-bad654a508b3 req-5c61ac36-7604-49ea-b651-a6c0436ad763 service nova] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Updating instance_info_cache with network_info: [{"id": "49b9f908-9366-4318-b2d3-c408e404a9df", "address": "fa:16:3e:82:b4:d8", "network": {"id": "49776280-f4e5-4f22-98fd-a6e6ff70c80f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1268791360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41ca5f8268434d329deec67fb2867b59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap49b9f908-93", "ovs_interfaceid": "49b9f908-9366-4318-b2d3-c408e404a9df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.378649] env[63372]: DEBUG nova.compute.utils [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 777.423009] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024014, 'name': CloneVM_Task, 'duration_secs': 1.725827} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.423463] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Created linked-clone VM from snapshot [ 777.424400] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3011033a-ddfc-4bd2-8c4c-8e6cdb77caad {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.432295] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Uploading image ec1d25a7-8ac1-4e60-bb42-a9bea2ae617a {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 777.460777] env[63372]: DEBUG oslo_vmware.rw_handles [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 777.460777] env[63372]: value = "vm-227343" [ 777.460777] env[63372]: _type = "VirtualMachine" [ 777.460777] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 777.461472] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-6850a8f0-1eb7-43d4-9279-e7789f9dcb4e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.471679] env[63372]: DEBUG oslo_vmware.rw_handles [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lease: (returnval){ [ 777.471679] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52094d8a-193b-0d1f-7a6a-b8a905fa51ea" [ 777.471679] env[63372]: _type = "HttpNfcLease" [ 777.471679] env[63372]: } obtained for exporting VM: (result){ [ 777.471679] env[63372]: value = "vm-227343" [ 777.471679] env[63372]: _type = "VirtualMachine" [ 777.471679] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 777.471679] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the lease: (returnval){ [ 777.471679] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52094d8a-193b-0d1f-7a6a-b8a905fa51ea" [ 777.471679] env[63372]: _type = "HttpNfcLease" [ 777.471679] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 777.477563] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 777.477563] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52094d8a-193b-0d1f-7a6a-b8a905fa51ea" [ 777.477563] env[63372]: _type = "HttpNfcLease" [ 777.477563] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 777.542543] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f2f691-49a2-b052-b8f6-5511ff184cca, 'name': SearchDatastore_Task, 'duration_secs': 0.013405} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.542952] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.543276] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 777.543606] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.543834] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.543946] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 777.544304] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-47b94711-c19f-4651-863d-ecfb49abc943 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.552849] env[63372]: DEBUG nova.network.neutron [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Successfully updated port: 116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 777.555241] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 777.555540] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 777.556655] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-759e15b7-132b-440c-a7b0-734f30c01251 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.565896] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 777.565896] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b32aa7-669b-c92d-b6df-6e4aff835bd4" [ 777.565896] env[63372]: _type = "Task" [ 777.565896] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.569892] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024021, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.578268] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b32aa7-669b-c92d-b6df-6e4aff835bd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.635100] env[63372]: DEBUG nova.compute.utils [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 777.637051] env[63372]: DEBUG nova.compute.manager [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Not allocating networking since 'none' was specified. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 777.662635] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.721305] env[63372]: DEBUG oslo_vmware.api [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024022, 'name': PowerOnVM_Task, 'duration_secs': 0.551346} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.721679] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 777.721887] env[63372]: INFO nova.compute.manager [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Took 11.45 seconds to spawn the instance on the hypervisor. [ 777.722091] env[63372]: DEBUG nova.compute.manager [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 777.723248] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ea9f8a-37e0-4c15-aa46-5da79c0de006 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.818698] env[63372]: DEBUG oslo_concurrency.lockutils [req-dadd4c5c-e765-49dc-9b87-bad654a508b3 req-5c61ac36-7604-49ea-b651-a6c0436ad763 service nova] Releasing lock "refresh_cache-bad91d77-c7ee-4572-b1ed-068b2a55233c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.881730] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "ac90a156-be00-4f62-a76e-e08914531167" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.980099] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 777.980099] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52094d8a-193b-0d1f-7a6a-b8a905fa51ea" [ 777.980099] env[63372]: _type = "HttpNfcLease" [ 777.980099] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 777.980408] env[63372]: DEBUG oslo_vmware.rw_handles [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 777.980408] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52094d8a-193b-0d1f-7a6a-b8a905fa51ea" [ 777.980408] env[63372]: _type = "HttpNfcLease" [ 777.980408] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 777.981163] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-480aaf4b-eabe-409e-bb3b-d7477f2dd792 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.988993] env[63372]: DEBUG oslo_vmware.rw_handles [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b2d8ec-be36-d0c8-f291-0214a6c994dc/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 777.989398] env[63372]: DEBUG oslo_vmware.rw_handles [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b2d8ec-be36-d0c8-f291-0214a6c994dc/disk-0.vmdk for reading. {{(pid=63372) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 778.057292] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "refresh_cache-cbd55518-a3b2-4636-ba43-c279fdd0bf8d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.057455] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired lock "refresh_cache-cbd55518-a3b2-4636-ba43-c279fdd0bf8d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.057612] env[63372]: DEBUG nova.network.neutron [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 778.069227] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024021, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.079897] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b32aa7-669b-c92d-b6df-6e4aff835bd4, 'name': SearchDatastore_Task, 'duration_secs': 0.013902} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.080763] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f17c6793-4b7d-4d31-9fb8-62a3046e2d74 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.086298] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 778.086298] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f1ddc2-830f-f5da-4968-afd804af15a1" [ 778.086298] env[63372]: _type = "Task" [ 778.086298] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.094900] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f1ddc2-830f-f5da-4968-afd804af15a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.097689] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-272d9cab-56dc-4765-90f4-82808242f0f8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.142244] env[63372]: DEBUG nova.compute.manager [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 778.254502] env[63372]: INFO nova.compute.manager [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Took 47.46 seconds to build instance. [ 778.575458] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024021, 'name': ReconfigVM_Task, 'duration_secs': 1.519456} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.575912] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Reconfigured VM instance instance-0000002f to attach disk [datastore2] 8283b736-ad02-4082-97b7-561bd5c5da93/8283b736-ad02-4082-97b7-561bd5c5da93.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 778.576294] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance '8283b736-ad02-4082-97b7-561bd5c5da93' progress to 50 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 778.587166] env[63372]: DEBUG nova.compute.manager [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Received event network-changed-29995495-a43f-4be1-b907-c88d9fe78dcc {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.587622] env[63372]: DEBUG nova.compute.manager [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Refreshing instance network info cache due to event network-changed-29995495-a43f-4be1-b907-c88d9fe78dcc. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 778.587622] env[63372]: DEBUG oslo_concurrency.lockutils [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] Acquiring lock "refresh_cache-69a107d0-80c7-42e5-b514-b4273e1a3359" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.587728] env[63372]: DEBUG oslo_concurrency.lockutils [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] Acquired lock "refresh_cache-69a107d0-80c7-42e5-b514-b4273e1a3359" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.587894] env[63372]: DEBUG nova.network.neutron [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Refreshing network info cache for port 29995495-a43f-4be1-b907-c88d9fe78dcc {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 778.603196] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f1ddc2-830f-f5da-4968-afd804af15a1, 'name': SearchDatastore_Task, 'duration_secs': 0.01191} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.604345] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.604724] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] bad91d77-c7ee-4572-b1ed-068b2a55233c/bad91d77-c7ee-4572-b1ed-068b2a55233c.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 778.606406] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-60cf25e1-49af-4221-a33c-a1fbeb633860 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.613418] env[63372]: DEBUG nova.network.neutron [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.623708] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 778.623708] env[63372]: value = "task-1024024" [ 778.623708] env[63372]: _type = "Task" [ 778.623708] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.639050] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024024, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.757250] env[63372]: DEBUG oslo_concurrency.lockutils [None req-50a325c4-dc54-4da2-8a59-4fea31d0f029 tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "febeca1c-3370-4c74-8bd9-efad313df1e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.107s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.763024] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3592ed74-93bc-4638-ad35-91a935708d37 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.776493] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2a43e69-e296-4a96-bd74-1a744f9779d2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.817058] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e64c61-fa58-4349-9047-e32b7d9553e0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.825778] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40715bdf-61fd-48c2-a93c-0f3f1100bd3a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.844686] env[63372]: DEBUG nova.compute.provider_tree [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.863637] env[63372]: DEBUG nova.network.neutron [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Updating instance_info_cache with network_info: [{"id": "116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee", "address": "fa:16:3e:70:f7:8a", "network": {"id": "49776280-f4e5-4f22-98fd-a6e6ff70c80f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1268791360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41ca5f8268434d329deec67fb2867b59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap116aaf8a-12", "ovs_interfaceid": "116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.973709] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "ac90a156-be00-4f62-a76e-e08914531167" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.974196] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "ac90a156-be00-4f62-a76e-e08914531167" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.974668] env[63372]: INFO nova.compute.manager [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Attaching volume e6717bc1-7823-4b07-889e-fd58898dee2a to /dev/sdb [ 779.025462] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0759c390-088e-4599-9391-19203d6bc02f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.036498] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9423558c-2b0f-4dab-96c8-626f2768a3ce {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.056335] env[63372]: DEBUG nova.virt.block_device [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Updating existing volume attachment record: c695294d-6ad4-46c9-bda8-65e9b1e16528 {{(pid=63372) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 779.086739] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-403608bb-83cf-4bd2-bfe3-6d5236e615ba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.120259] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ee527f-ffc9-4ba1-817f-2975665c4763 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.146882] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "febeca1c-3370-4c74-8bd9-efad313df1e2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.147333] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "febeca1c-3370-4c74-8bd9-efad313df1e2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.147721] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "febeca1c-3370-4c74-8bd9-efad313df1e2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.147950] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "febeca1c-3370-4c74-8bd9-efad313df1e2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.148206] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "febeca1c-3370-4c74-8bd9-efad313df1e2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.150773] env[63372]: INFO nova.compute.manager [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Terminating instance [ 779.153274] env[63372]: DEBUG nova.compute.manager [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 779.153733] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 779.154199] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance '8283b736-ad02-4082-97b7-561bd5c5da93' progress to 67 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 779.161484] env[63372]: DEBUG nova.compute.manager [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 779.168186] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac60473-b09f-4597-9e0e-cf715f897fb8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.179806] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024024, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.182931] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 779.184399] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7cad078-ea47-413b-9f4e-02829232ab8a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.195272] env[63372]: DEBUG oslo_vmware.api [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 779.195272] env[63372]: value = "task-1024025" [ 779.195272] env[63372]: _type = "Task" [ 779.195272] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.210451] env[63372]: DEBUG oslo_vmware.api [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024025, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.270371] env[63372]: DEBUG nova.compute.manager [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 779.346585] env[63372]: DEBUG nova.scheduler.client.report [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.367803] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Releasing lock "refresh_cache-cbd55518-a3b2-4636-ba43-c279fdd0bf8d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.368482] env[63372]: DEBUG nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Instance network_info: |[{"id": "116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee", "address": "fa:16:3e:70:f7:8a", "network": {"id": "49776280-f4e5-4f22-98fd-a6e6ff70c80f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1268791360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41ca5f8268434d329deec67fb2867b59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap116aaf8a-12", "ovs_interfaceid": "116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 779.369193] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:f7:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffcecdaa-a7b8-49fc-9371-dbdb7744688e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 779.379745] env[63372]: DEBUG oslo.service.loopingcall [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.379745] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 779.379745] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fbe51b98-1422-4c8f-b13b-6e5fe0685278 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.398554] env[63372]: DEBUG oslo_concurrency.lockutils [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Acquiring lock "4c9eb955-3bed-4b26-866f-b2a876b835dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.398896] env[63372]: DEBUG oslo_concurrency.lockutils [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Lock "4c9eb955-3bed-4b26-866f-b2a876b835dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.399138] env[63372]: DEBUG oslo_concurrency.lockutils [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Acquiring lock "4c9eb955-3bed-4b26-866f-b2a876b835dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.399505] env[63372]: DEBUG oslo_concurrency.lockutils [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Lock "4c9eb955-3bed-4b26-866f-b2a876b835dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.399781] env[63372]: DEBUG oslo_concurrency.lockutils [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Lock "4c9eb955-3bed-4b26-866f-b2a876b835dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.407086] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 779.407086] env[63372]: value = "task-1024029" [ 779.407086] env[63372]: _type = "Task" [ 779.407086] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.407858] env[63372]: INFO nova.compute.manager [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Terminating instance [ 779.414023] env[63372]: DEBUG nova.compute.manager [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 779.414023] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 779.417719] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93869bc5-2816-4420-9f04-3ae0bb519680 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.427106] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024029, 'name': CreateVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.430736] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 779.431171] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4731243a-4627-4677-8708-8d5d171dc6df {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.440118] env[63372]: DEBUG oslo_vmware.api [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Waiting for the task: (returnval){ [ 779.440118] env[63372]: value = "task-1024030" [ 779.440118] env[63372]: _type = "Task" [ 779.440118] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.454036] env[63372]: DEBUG oslo_vmware.api [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1024030, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.458628] env[63372]: DEBUG nova.network.neutron [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Updated VIF entry in instance network info cache for port 29995495-a43f-4be1-b907-c88d9fe78dcc. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 779.459375] env[63372]: DEBUG nova.network.neutron [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Updating instance_info_cache with network_info: [{"id": "29995495-a43f-4be1-b907-c88d9fe78dcc", "address": "fa:16:3e:cb:ee:42", "network": {"id": "3ca584f0-a422-4ca2-bdc9-033462b8580e", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-175791853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "541cfc7f0fba40678e1f67d677d22df2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c6a4836-66dc-4e43-982b-f8fcd3f9989a", "external-id": "nsx-vlan-transportzone-635", "segmentation_id": 635, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29995495-a4", "ovs_interfaceid": "29995495-a43f-4be1-b907-c88d9fe78dcc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.641068] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024024, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.710484} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.641563] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] bad91d77-c7ee-4572-b1ed-068b2a55233c/bad91d77-c7ee-4572-b1ed-068b2a55233c.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 779.641824] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 779.642195] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a99ab9d4-613c-4e0d-8dc7-07e329391957 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.652703] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 779.652703] env[63372]: value = "task-1024031" [ 779.652703] env[63372]: _type = "Task" [ 779.652703] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.663679] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024031, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.709292] env[63372]: DEBUG oslo_vmware.api [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024025, 'name': PowerOffVM_Task, 'duration_secs': 0.293591} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.709783] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 779.710064] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 779.710484] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5c0685e6-0b83-48e5-b105-bcb2fc9b5167 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.719993] env[63372]: DEBUG nova.network.neutron [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Port 874f14b1-6cc3-4701-8a08-35940fc9bd9b binding to destination host cpu-1 is already ACTIVE {{(pid=63372) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 779.797969] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.855018] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.732s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.855769] env[63372]: DEBUG nova.compute.manager [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 779.860202] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.384s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.862772] env[63372]: INFO nova.compute.claims [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 779.867908] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 779.868457] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 779.868799] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Deleting the datastore file [datastore2] febeca1c-3370-4c74-8bd9-efad313df1e2 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 779.869464] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-237d7f64-50d7-4510-8181-513f3ffa6e0b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.881061] env[63372]: DEBUG oslo_vmware.api [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for the task: (returnval){ [ 779.881061] env[63372]: value = "task-1024033" [ 779.881061] env[63372]: _type = "Task" [ 779.881061] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.891741] env[63372]: DEBUG oslo_vmware.api [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024033, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.921860] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024029, 'name': CreateVM_Task, 'duration_secs': 0.504413} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.922112] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 779.923023] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.923294] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.923885] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 779.924490] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fed4bff-1b12-4929-999f-fff6a0634af3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.930890] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 779.930890] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]522974eb-6a64-4109-d4f5-b6df4773fbdd" [ 779.930890] env[63372]: _type = "Task" [ 779.930890] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.942174] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522974eb-6a64-4109-d4f5-b6df4773fbdd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.952529] env[63372]: DEBUG oslo_vmware.api [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1024030, 'name': PowerOffVM_Task, 'duration_secs': 0.267704} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.952929] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 779.953194] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 779.953472] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-68cf2b99-b40b-4b25-a15a-2afc32c17909 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.962729] env[63372]: DEBUG oslo_concurrency.lockutils [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] Releasing lock "refresh_cache-69a107d0-80c7-42e5-b514-b4273e1a3359" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.963206] env[63372]: DEBUG nova.compute.manager [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Received event network-changed-116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 779.963454] env[63372]: DEBUG nova.compute.manager [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Refreshing instance network info cache due to event network-changed-116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 779.963701] env[63372]: DEBUG oslo_concurrency.lockutils [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] Acquiring lock "refresh_cache-cbd55518-a3b2-4636-ba43-c279fdd0bf8d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.963891] env[63372]: DEBUG oslo_concurrency.lockutils [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] Acquired lock "refresh_cache-cbd55518-a3b2-4636-ba43-c279fdd0bf8d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.964269] env[63372]: DEBUG nova.network.neutron [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Refreshing network info cache for port 116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 780.032861] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 780.033166] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 780.033392] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Deleting the datastore file [datastore2] 4c9eb955-3bed-4b26-866f-b2a876b835dd {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 780.033753] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fb836848-7e98-45db-9d9e-9efdf193fd51 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.041477] env[63372]: DEBUG oslo_vmware.api [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Waiting for the task: (returnval){ [ 780.041477] env[63372]: value = "task-1024035" [ 780.041477] env[63372]: _type = "Task" [ 780.041477] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.050522] env[63372]: DEBUG oslo_vmware.api [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1024035, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.163637] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024031, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087128} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.163888] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 780.164716] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72719445-8c8e-4dc1-a566-89c9dd21e4d7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.190606] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] bad91d77-c7ee-4572-b1ed-068b2a55233c/bad91d77-c7ee-4572-b1ed-068b2a55233c.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 780.190957] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-644d1f86-2304-497d-85e9-d54d1ecb44b8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.215420] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 780.215420] env[63372]: value = "task-1024036" [ 780.215420] env[63372]: _type = "Task" [ 780.215420] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.232308] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024036, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.368465] env[63372]: DEBUG nova.compute.utils [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.370055] env[63372]: DEBUG nova.compute.manager [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Not allocating networking since 'none' was specified. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 780.393322] env[63372]: DEBUG oslo_vmware.api [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Task: {'id': task-1024033, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.329901} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.393627] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 780.393833] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 780.394021] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 780.394223] env[63372]: INFO nova.compute.manager [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Took 1.24 seconds to destroy the instance on the hypervisor. [ 780.394476] env[63372]: DEBUG oslo.service.loopingcall [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.394684] env[63372]: DEBUG nova.compute.manager [-] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.394798] env[63372]: DEBUG nova.network.neutron [-] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 780.442936] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522974eb-6a64-4109-d4f5-b6df4773fbdd, 'name': SearchDatastore_Task, 'duration_secs': 0.01446} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.443417] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.443813] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 780.444411] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.444411] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.444777] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 780.445284] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f20d302b-58d0-4f04-adb3-64daa782cdc3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.458243] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 780.458243] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 780.459331] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b69eebeb-b9d3-4ea8-bc64-b99272c3853d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.470111] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 780.470111] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]522232a0-471b-06e2-03ad-b3c9558ba82c" [ 780.470111] env[63372]: _type = "Task" [ 780.470111] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.480500] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522232a0-471b-06e2-03ad-b3c9558ba82c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.552580] env[63372]: DEBUG oslo_vmware.api [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Task: {'id': task-1024035, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.301022} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.553044] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 780.553267] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 780.553455] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 780.553638] env[63372]: INFO nova.compute.manager [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Took 1.14 seconds to destroy the instance on the hypervisor. [ 780.553880] env[63372]: DEBUG oslo.service.loopingcall [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.554093] env[63372]: DEBUG nova.compute.manager [-] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.554182] env[63372]: DEBUG nova.network.neutron [-] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 780.727424] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024036, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.748777] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "8283b736-ad02-4082-97b7-561bd5c5da93-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.749035] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "8283b736-ad02-4082-97b7-561bd5c5da93-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.749221] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "8283b736-ad02-4082-97b7-561bd5c5da93-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.817463] env[63372]: DEBUG nova.network.neutron [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Updated VIF entry in instance network info cache for port 116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 780.817695] env[63372]: DEBUG nova.network.neutron [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Updating instance_info_cache with network_info: [{"id": "116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee", "address": "fa:16:3e:70:f7:8a", "network": {"id": "49776280-f4e5-4f22-98fd-a6e6ff70c80f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1268791360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41ca5f8268434d329deec67fb2867b59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap116aaf8a-12", "ovs_interfaceid": "116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.874279] env[63372]: DEBUG nova.compute.manager [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 780.983579] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522232a0-471b-06e2-03ad-b3c9558ba82c, 'name': SearchDatastore_Task, 'duration_secs': 0.014535} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.984506] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7900cf2-ebc1-4288-97f1-6c75f71b635d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.997058] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 780.997058] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52061cf0-8109-6962-3177-88aba8d1ed30" [ 780.997058] env[63372]: _type = "Task" [ 780.997058] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.008106] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52061cf0-8109-6962-3177-88aba8d1ed30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.229840] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024036, 'name': ReconfigVM_Task, 'duration_secs': 0.552602} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.233092] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Reconfigured VM instance instance-0000003a to attach disk [datastore2] bad91d77-c7ee-4572-b1ed-068b2a55233c/bad91d77-c7ee-4572-b1ed-068b2a55233c.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.234174] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e8387dc5-684b-4e4a-b56c-200387cb345e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.256498] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 781.256498] env[63372]: value = "task-1024037" [ 781.256498] env[63372]: _type = "Task" [ 781.256498] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.270693] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024037, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.320713] env[63372]: DEBUG oslo_concurrency.lockutils [req-5c12c1af-8f9f-4179-b927-d518732018d7 req-bcb2ea5d-c2d5-45e4-b065-9e294eee695a service nova] Releasing lock "refresh_cache-cbd55518-a3b2-4636-ba43-c279fdd0bf8d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.335202] env[63372]: DEBUG nova.network.neutron [-] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.448304] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d139b4ea-28c8-4725-8e57-b595fda1d53d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.457512] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef305f3-9a1a-43cf-8887-0b842a4ea1e2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.491628] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb5f4fd-62b7-4811-b2b7-9e2a3f4458b0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.503771] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ad289a-28c7-45cb-9c45-4c3a1a026ef7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.519989] env[63372]: DEBUG nova.compute.provider_tree [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.525107] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52061cf0-8109-6962-3177-88aba8d1ed30, 'name': SearchDatastore_Task, 'duration_secs': 0.015961} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.525691] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.525978] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] cbd55518-a3b2-4636-ba43-c279fdd0bf8d/cbd55518-a3b2-4636-ba43-c279fdd0bf8d.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 781.526259] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db62037c-e523-4d3a-b9dd-bace5ac9e453 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.534252] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 781.534252] env[63372]: value = "task-1024039" [ 781.534252] env[63372]: _type = "Task" [ 781.534252] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.543286] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024039, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.545739] env[63372]: DEBUG nova.network.neutron [-] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.769124] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024037, 'name': Rename_Task, 'duration_secs': 0.24324} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.769448] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 781.769724] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fcb488b6-c525-4056-a282-925c5b70e068 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.778339] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 781.778339] env[63372]: value = "task-1024040" [ 781.778339] env[63372]: _type = "Task" [ 781.778339] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.791680] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024040, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.817688] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.818014] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquired lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.818701] env[63372]: DEBUG nova.network.neutron [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 781.837742] env[63372]: INFO nova.compute.manager [-] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Took 1.28 seconds to deallocate network for instance. [ 781.889441] env[63372]: DEBUG nova.compute.manager [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 782.027545] env[63372]: DEBUG nova.scheduler.client.report [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.047679] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024039, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.048212] env[63372]: INFO nova.compute.manager [-] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Took 1.65 seconds to deallocate network for instance. [ 782.291364] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024040, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.345718] env[63372]: DEBUG oslo_concurrency.lockutils [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.533540] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.673s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.534106] env[63372]: DEBUG nova.compute.manager [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 782.536724] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.724s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.538469] env[63372]: INFO nova.compute.claims [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.551477] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024039, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546919} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.551718] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] cbd55518-a3b2-4636-ba43-c279fdd0bf8d/cbd55518-a3b2-4636-ba43-c279fdd0bf8d.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 782.551955] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 782.552261] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d8353d72-48eb-4e33-91d3-59e8f8b7ce25 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.555012] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.561993] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 782.561993] env[63372]: value = "task-1024041" [ 782.561993] env[63372]: _type = "Task" [ 782.561993] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.563104] env[63372]: DEBUG nova.network.neutron [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance_info_cache with network_info: [{"id": "874f14b1-6cc3-4701-8a08-35940fc9bd9b", "address": "fa:16:3e:57:46:8e", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.24", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap874f14b1-6c", "ovs_interfaceid": "874f14b1-6cc3-4701-8a08-35940fc9bd9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.575273] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024041, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.791575] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024040, 'name': PowerOnVM_Task, 'duration_secs': 0.70417} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.791978] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 782.792207] env[63372]: INFO nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Took 9.15 seconds to spawn the instance on the hypervisor. [ 782.792396] env[63372]: DEBUG nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.793360] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ea978b-6bb4-4650-b1fb-98eccb0f6bf1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.044089] env[63372]: DEBUG nova.compute.utils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 783.045093] env[63372]: DEBUG nova.compute.manager [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 783.045301] env[63372]: DEBUG nova.network.neutron [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 783.070035] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Releasing lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.080354] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024041, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088212} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.080661] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 783.081510] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-085af55d-499c-4389-86bf-91829077aba9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.089012] env[63372]: DEBUG nova.policy [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d247cdc706e4f48a38c6ed1852b7a85', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1cc87c31eb77441daeec0fad5bb1ce87', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 783.108685] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] cbd55518-a3b2-4636-ba43-c279fdd0bf8d/cbd55518-a3b2-4636-ba43-c279fdd0bf8d.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 783.109328] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68650891-9d8d-4c8a-8c57-4c9d7de66951 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.132604] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 783.132604] env[63372]: value = "task-1024042" [ 783.132604] env[63372]: _type = "Task" [ 783.132604] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.143023] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024042, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.314123] env[63372]: INFO nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Took 44.46 seconds to build instance. [ 783.361452] env[63372]: DEBUG nova.network.neutron [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Successfully created port: 0e3060b6-6d21-4a2d-8e8a-ae48e64ee869 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 783.548672] env[63372]: DEBUG nova.compute.manager [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 783.596143] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12583127-6a7a-4fb1-ad09-692d511b43a1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.617345] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Volume attach. Driver type: vmdk {{(pid=63372) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 783.617619] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227348', 'volume_id': 'e6717bc1-7823-4b07-889e-fd58898dee2a', 'name': 'volume-e6717bc1-7823-4b07-889e-fd58898dee2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ac90a156-be00-4f62-a76e-e08914531167', 'attached_at': '', 'detached_at': '', 'volume_id': 'e6717bc1-7823-4b07-889e-fd58898dee2a', 'serial': 'e6717bc1-7823-4b07-889e-fd58898dee2a'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 783.621530] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b20002d-5383-4c0e-9122-d671741adbb3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.624852] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa42c6b-4106-4e2d-8331-1b900a1b9bbf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.633187] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance '8283b736-ad02-4082-97b7-561bd5c5da93' progress to 83 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 783.657799] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb47129-ef33-4f2d-a763-38431c2568c4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.666997] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024042, 'name': ReconfigVM_Task, 'duration_secs': 0.422493} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.680598] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Reconfigured VM instance instance-0000003b to attach disk [datastore1] cbd55518-a3b2-4636-ba43-c279fdd0bf8d/cbd55518-a3b2-4636-ba43-c279fdd0bf8d.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 783.689569] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Reconfiguring VM instance instance-00000021 to attach disk [datastore1] volume-e6717bc1-7823-4b07-889e-fd58898dee2a/volume-e6717bc1-7823-4b07-889e-fd58898dee2a.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 783.692532] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61ef249c-911b-418e-8c94-af045d1704e9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.694570] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c0793a2-7493-43c8-bc9c-5e2fecea1f81 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.715670] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 783.715670] env[63372]: value = "task-1024043" [ 783.715670] env[63372]: _type = "Task" [ 783.715670] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.717149] env[63372]: DEBUG oslo_vmware.api [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 783.717149] env[63372]: value = "task-1024044" [ 783.717149] env[63372]: _type = "Task" [ 783.717149] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.733286] env[63372]: DEBUG oslo_vmware.api [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024044, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.736848] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024043, 'name': Rename_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.816895] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "bad91d77-c7ee-4572-b1ed-068b2a55233c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.194s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.102234] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d30dd0d-e2b1-420d-bc75-9343a32cfaf4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.111398] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e811f26-39bb-4c1b-b5fb-93e0d7422426 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.146564] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6703a8-50ba-427f-84a4-1a52ef2d0d47 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.156117] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 784.156494] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cbe00d75-a85c-44e6-b533-bc9e1f85a1b7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.159334] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7090cd0-ba2c-4c87-8510-e3d9008378de {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.175955] env[63372]: DEBUG nova.compute.provider_tree [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.179032] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 784.179032] env[63372]: value = "task-1024045" [ 784.179032] env[63372]: _type = "Task" [ 784.179032] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.188045] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024045, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.232478] env[63372]: DEBUG oslo_vmware.api [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024044, 'name': ReconfigVM_Task, 'duration_secs': 0.429115} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.236621] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Reconfigured VM instance instance-00000021 to attach disk [datastore1] volume-e6717bc1-7823-4b07-889e-fd58898dee2a/volume-e6717bc1-7823-4b07-889e-fd58898dee2a.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.242290] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024043, 'name': Rename_Task, 'duration_secs': 0.203915} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.242730] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5757a2e-9619-4c1e-916c-1e564252d1f7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.255540] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 784.256462] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07f494f3-a5a8-4f42-bbf5-433fde2948f4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.265588] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 784.265588] env[63372]: value = "task-1024046" [ 784.265588] env[63372]: _type = "Task" [ 784.265588] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.267174] env[63372]: DEBUG oslo_vmware.api [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 784.267174] env[63372]: value = "task-1024047" [ 784.267174] env[63372]: _type = "Task" [ 784.267174] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.279741] env[63372]: DEBUG oslo_vmware.api [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024047, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.282962] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024046, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.322190] env[63372]: DEBUG nova.compute.manager [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 784.564112] env[63372]: DEBUG nova.compute.manager [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 784.680538] env[63372]: DEBUG nova.scheduler.client.report [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.694435] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024045, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.782917] env[63372]: DEBUG oslo_vmware.api [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024047, 'name': ReconfigVM_Task, 'duration_secs': 0.169886} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.786803] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227348', 'volume_id': 'e6717bc1-7823-4b07-889e-fd58898dee2a', 'name': 'volume-e6717bc1-7823-4b07-889e-fd58898dee2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ac90a156-be00-4f62-a76e-e08914531167', 'attached_at': '', 'detached_at': '', 'volume_id': 'e6717bc1-7823-4b07-889e-fd58898dee2a', 'serial': 'e6717bc1-7823-4b07-889e-fd58898dee2a'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 784.788809] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024046, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.981707] env[63372]: DEBUG nova.network.neutron [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Successfully updated port: 0e3060b6-6d21-4a2d-8e8a-ae48e64ee869 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 784.985409] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.192122] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.655s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.193078] env[63372]: DEBUG nova.compute.manager [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.197369] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.621s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.199358] env[63372]: INFO nova.compute.claims [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 785.211888] env[63372]: DEBUG oslo_vmware.api [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024045, 'name': PowerOnVM_Task, 'duration_secs': 1.033468} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.212671] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 785.212671] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9de07e4a-ea94-4061-9d5c-6ac907a31704 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance '8283b736-ad02-4082-97b7-561bd5c5da93' progress to 100 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 785.279224] env[63372]: DEBUG oslo_vmware.api [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024046, 'name': PowerOnVM_Task, 'duration_secs': 0.96687} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.280292] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 785.280525] env[63372]: INFO nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Took 8.87 seconds to spawn the instance on the hypervisor. [ 785.280709] env[63372]: DEBUG nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 785.281556] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4edd501f-dc1a-4e95-81b1-4c7deca131fd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.486113] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "refresh_cache-99f901a6-9bb3-4403-af0c-c8900f655cb3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.486252] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquired lock "refresh_cache-99f901a6-9bb3-4403-af0c-c8900f655cb3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.486407] env[63372]: DEBUG nova.network.neutron [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 785.642690] env[63372]: DEBUG nova.virt.hardware [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 785.643226] env[63372]: DEBUG nova.virt.hardware [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 785.643552] env[63372]: DEBUG nova.virt.hardware [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 785.643857] env[63372]: DEBUG nova.virt.hardware [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 785.644122] env[63372]: DEBUG nova.virt.hardware [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 785.644384] env[63372]: DEBUG nova.virt.hardware [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 785.645076] env[63372]: DEBUG nova.virt.hardware [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 785.645076] env[63372]: DEBUG nova.virt.hardware [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 785.645076] env[63372]: DEBUG nova.virt.hardware [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 785.645304] env[63372]: DEBUG nova.virt.hardware [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 785.645537] env[63372]: DEBUG nova.virt.hardware [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 785.647837] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119e2e08-cf1c-4708-bb3a-9365a264d9be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.658198] env[63372]: DEBUG nova.virt.hardware [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 785.658581] env[63372]: DEBUG nova.virt.hardware [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 785.658661] env[63372]: DEBUG nova.virt.hardware [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 785.658815] env[63372]: DEBUG nova.virt.hardware [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 785.658963] env[63372]: DEBUG nova.virt.hardware [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 785.659124] env[63372]: DEBUG nova.virt.hardware [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 785.659331] env[63372]: DEBUG nova.virt.hardware [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 785.659494] env[63372]: DEBUG nova.virt.hardware [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 785.659664] env[63372]: DEBUG nova.virt.hardware [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 785.659824] env[63372]: DEBUG nova.virt.hardware [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 785.660012] env[63372]: DEBUG nova.virt.hardware [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 785.661274] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a798a6ab-9000-4486-93f4-475637b2e472 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.666767] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88c64152-6678-4a8d-a268-4e43e32203a8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.676803] env[63372]: DEBUG nova.virt.hardware [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 785.677039] env[63372]: DEBUG nova.virt.hardware [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 785.677239] env[63372]: DEBUG nova.virt.hardware [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 785.677431] env[63372]: DEBUG nova.virt.hardware [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 785.678566] env[63372]: DEBUG nova.virt.hardware [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 785.678566] env[63372]: DEBUG nova.virt.hardware [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 785.678566] env[63372]: DEBUG nova.virt.hardware [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 785.678566] env[63372]: DEBUG nova.virt.hardware [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 785.678566] env[63372]: DEBUG nova.virt.hardware [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 785.678566] env[63372]: DEBUG nova.virt.hardware [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 785.678842] env[63372]: DEBUG nova.virt.hardware [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 785.687378] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b368c0-05a1-4112-98ad-0efa48574e29 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.693069] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bdc7ea3-9f61-47d2-8c82-19b9e5f9ad10 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.695592] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Instance VIF info [] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 785.701882] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Creating folder: Project (32969ae92e89421c90a4cee15dc69614). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 785.702229] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2a0e29ff-8411-4cf7-a661-f434dd7785d4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.705643] env[63372]: DEBUG nova.compute.utils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 785.708235] env[63372]: DEBUG oslo_vmware.rw_handles [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526cdd31-0293-75ee-5fce-bdd2599d8960/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 785.711379] env[63372]: DEBUG nova.compute.manager [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 785.711609] env[63372]: DEBUG nova.network.neutron [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 785.714852] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90143f00-1409-4e51-a597-62a42a61eac7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.732029] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Instance VIF info [] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 785.737967] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Creating folder: Project (32969ae92e89421c90a4cee15dc69614). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 785.739659] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7ae1e660-1264-46b2-953a-5282ca1e768b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.744012] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5010b517-e3d1-4e33-9dc5-3b7a2ceeb1d2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.750951] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Created folder: Project (32969ae92e89421c90a4cee15dc69614) in parent group-v227230. [ 785.751167] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Creating folder: Instances. Parent ref: group-v227350. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 785.753677] env[63372]: DEBUG oslo_vmware.rw_handles [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526cdd31-0293-75ee-5fce-bdd2599d8960/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 785.753677] env[63372]: ERROR oslo_vmware.rw_handles [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526cdd31-0293-75ee-5fce-bdd2599d8960/disk-0.vmdk due to incomplete transfer. [ 785.753677] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f943226c-541f-402b-b93e-13477a82d3d4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.754507] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-db090178-535c-497a-ab90-597ce9780ddb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.767852] env[63372]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 785.767852] env[63372]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63372) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 785.768799] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Folder already exists: Project (32969ae92e89421c90a4cee15dc69614). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.768986] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Creating folder: Instances. Parent ref: group-v227350. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 785.769228] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e7615342-5fcc-4aef-ab39-ba7ea0ba8438 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.780921] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Created folder: Instances in parent group-v227350. [ 785.781280] env[63372]: DEBUG oslo.service.loopingcall [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.784614] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 785.784614] env[63372]: DEBUG oslo_vmware.rw_handles [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526cdd31-0293-75ee-5fce-bdd2599d8960/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 785.784614] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Uploaded image 442ec676-c4b4-4d9a-80ab-4267edc97c35 to the Glance image server {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 785.785590] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 785.789021] env[63372]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 785.789021] env[63372]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63372) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 785.789021] env[63372]: DEBUG nova.policy [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d247cdc706e4f48a38c6ed1852b7a85', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1cc87c31eb77441daeec0fad5bb1ce87', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 785.789799] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-77a0f18e-2f4f-47a4-9c8a-c616013e0e4c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.803026] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-40bac2cb-9e36-46bd-b5b4-46d6421d6d95 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.806294] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Folder already exists: Instances. Parent ref: group-v227350. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.806544] env[63372]: DEBUG oslo.service.loopingcall [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.810866] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6675697-1529-46be-b28a-398ff3060d18] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 785.819970] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b1c7490-373e-48b0-ae0e-e76892dcacfd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.834886] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 785.834886] env[63372]: value = "task-1024052" [ 785.834886] env[63372]: _type = "Task" [ 785.834886] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.835781] env[63372]: DEBUG nova.objects.instance [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lazy-loading 'flavor' on Instance uuid ac90a156-be00-4f62-a76e-e08914531167 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 785.837510] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.837510] env[63372]: value = "task-1024053" [ 785.837510] env[63372]: _type = "Task" [ 785.837510] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.837924] env[63372]: INFO nova.compute.manager [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Took 42.22 seconds to build instance. [ 785.847302] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.847302] env[63372]: value = "task-1024054" [ 785.847302] env[63372]: _type = "Task" [ 785.847302] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.859036] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024052, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.859036] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024053, 'name': CreateVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.868494] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024054, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.041098] env[63372]: DEBUG nova.network.neutron [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 786.219364] env[63372]: DEBUG nova.compute.manager [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 786.334702] env[63372]: DEBUG nova.compute.manager [req-3f89dfc0-fe5f-46d3-8375-7b42b66a7379 req-def1c11b-2c3f-4d0c-9833-4ad47650211f service nova] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Received event network-vif-deleted-f7d8a5f6-6c1f-43fd-9e49-67ebdb41e0ff {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.345029] env[63372]: DEBUG oslo_concurrency.lockutils [None req-df3ceebe-7f1d-449e-ac62-685cc3d6af7d tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "cbd55518-a3b2-4636-ba43-c279fdd0bf8d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.689s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.358685] env[63372]: DEBUG nova.network.neutron [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Updating instance_info_cache with network_info: [{"id": "0e3060b6-6d21-4a2d-8e8a-ae48e64ee869", "address": "fa:16:3e:81:de:69", "network": {"id": "8171761d-bd68-4efd-ad4f-e661badde0c4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1048168720-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1cc87c31eb77441daeec0fad5bb1ce87", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e3060b6-6d", "ovs_interfaceid": "0e3060b6-6d21-4a2d-8e8a-ae48e64ee869", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.359741] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cadf0cbf-6e74-412e-8814-402ee9f7e1c4 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "ac90a156-be00-4f62-a76e-e08914531167" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.386s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.365344] env[63372]: DEBUG nova.compute.manager [req-2aeb1f78-9177-411b-a673-fe5fb8f13cc8 req-3ecbe884-1d7a-4193-bedb-a417cb653517 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Received event network-vif-deleted-9249f3d1-43e4-4e00-94a1-d4ed0b20b171 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.367360] env[63372]: DEBUG nova.network.neutron [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Successfully created port: 5bf209d9-9f64-410a-b14f-19cd792d4f7c {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.386397] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024052, 'name': Destroy_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.386962] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024053, 'name': CreateVM_Task, 'duration_secs': 0.456885} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.387830] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024054, 'name': CreateVM_Task, 'duration_secs': 0.361243} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.391605] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 786.391605] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6675697-1529-46be-b28a-398ff3060d18] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 786.391605] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.391785] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.391969] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 786.392934] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.392934] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c5e377d-cd4e-458f-b96c-d51f01e2a007 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.400065] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 786.400065] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5251474c-6b93-1833-aa33-a1aa8e88bb30" [ 786.400065] env[63372]: _type = "Task" [ 786.400065] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.413255] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5251474c-6b93-1833-aa33-a1aa8e88bb30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.732032] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e7eeb9-6cc5-4492-bbf3-ffcdc615dedb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.743232] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c77b749d-33b8-49a4-b60b-bacc662334ad {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.794637] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b2d35c4-d2c4-449f-90df-357950fca5e6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.803100] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bddde42-c9d5-46b4-bac6-6e1cbd25e490 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.819877] env[63372]: DEBUG nova.compute.provider_tree [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.850761] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024052, 'name': Destroy_Task, 'duration_secs': 0.650305} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.851109] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Destroyed the VM [ 786.851341] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 786.851664] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-dd0dcdc1-c7c4-4ba8-b9a2-63f6056a9c42 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.860259] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 786.860259] env[63372]: value = "task-1024055" [ 786.860259] env[63372]: _type = "Task" [ 786.860259] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.869663] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Releasing lock "refresh_cache-99f901a6-9bb3-4403-af0c-c8900f655cb3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.869966] env[63372]: DEBUG nova.compute.manager [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Instance network_info: |[{"id": "0e3060b6-6d21-4a2d-8e8a-ae48e64ee869", "address": "fa:16:3e:81:de:69", "network": {"id": "8171761d-bd68-4efd-ad4f-e661badde0c4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1048168720-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1cc87c31eb77441daeec0fad5bb1ce87", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e3060b6-6d", "ovs_interfaceid": "0e3060b6-6d21-4a2d-8e8a-ae48e64ee869", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 786.870469] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024055, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.870816] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:de:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0e3060b6-6d21-4a2d-8e8a-ae48e64ee869', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 786.878515] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Creating folder: Project (1cc87c31eb77441daeec0fad5bb1ce87). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 786.878796] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7558ef4-1b63-49da-bd2d-7c40242bf77e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.890438] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Created folder: Project (1cc87c31eb77441daeec0fad5bb1ce87) in parent group-v227230. [ 786.890684] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Creating folder: Instances. Parent ref: group-v227354. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 786.890942] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9461d102-d162-4b20-bc56-ab5800263536 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.900391] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Created folder: Instances in parent group-v227354. [ 786.900610] env[63372]: DEBUG oslo.service.loopingcall [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.900872] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 786.906023] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b09f111a-20fd-47cd-86a2-50179d801ed6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.929790] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5251474c-6b93-1833-aa33-a1aa8e88bb30, 'name': SearchDatastore_Task, 'duration_secs': 0.019829} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.931366] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.931629] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 786.931883] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.932064] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.932268] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 786.932527] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 786.932527] env[63372]: value = "task-1024058" [ 786.932527] env[63372]: _type = "Task" [ 786.932527] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.932754] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.936020] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 786.936020] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cb8b058a-9424-4a8a-aff2-94550861935d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.936020] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ba4670f-8ff8-4599-b73a-d0af0264898b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.943897] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 786.943897] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520f3db2-5a55-ce1c-013e-1dc5197c9c5b" [ 786.943897] env[63372]: _type = "Task" [ 786.943897] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.951300] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024058, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.957165] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520f3db2-5a55-ce1c-013e-1dc5197c9c5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.958428] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 786.958738] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 786.959391] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4d448fb-141e-4ae7-8411-612e61e79ac8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.965510] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 786.965510] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52aad9b5-248b-bc42-a84f-95a3ec9fb53d" [ 786.965510] env[63372]: _type = "Task" [ 786.965510] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.974597] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52aad9b5-248b-bc42-a84f-95a3ec9fb53d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.139034] env[63372]: DEBUG oslo_concurrency.lockutils [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Acquiring lock "7a75a51d-b804-453b-ba7b-fda6c4931802" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.139150] env[63372]: DEBUG oslo_concurrency.lockutils [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Lock "7a75a51d-b804-453b-ba7b-fda6c4931802" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.139281] env[63372]: DEBUG oslo_concurrency.lockutils [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Acquiring lock "7a75a51d-b804-453b-ba7b-fda6c4931802-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.139469] env[63372]: DEBUG oslo_concurrency.lockutils [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Lock "7a75a51d-b804-453b-ba7b-fda6c4931802-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.139652] env[63372]: DEBUG oslo_concurrency.lockutils [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Lock "7a75a51d-b804-453b-ba7b-fda6c4931802-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.144535] env[63372]: INFO nova.compute.manager [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Terminating instance [ 787.149193] env[63372]: DEBUG nova.compute.manager [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 787.149307] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 787.150306] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c299b100-6e72-4e3d-ade6-5ee8de0f194d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.163337] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 787.166164] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7776159e-9405-4ffe-b6e9-5ededcfbe580 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.180401] env[63372]: DEBUG oslo_vmware.api [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Waiting for the task: (returnval){ [ 787.180401] env[63372]: value = "task-1024059" [ 787.180401] env[63372]: _type = "Task" [ 787.180401] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.197879] env[63372]: DEBUG oslo_vmware.api [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1024059, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.238041] env[63372]: DEBUG nova.compute.manager [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.265450] env[63372]: DEBUG nova.virt.hardware [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.265774] env[63372]: DEBUG nova.virt.hardware [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.265921] env[63372]: DEBUG nova.virt.hardware [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.266129] env[63372]: DEBUG nova.virt.hardware [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.266287] env[63372]: DEBUG nova.virt.hardware [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.266437] env[63372]: DEBUG nova.virt.hardware [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.266652] env[63372]: DEBUG nova.virt.hardware [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.266817] env[63372]: DEBUG nova.virt.hardware [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.267020] env[63372]: DEBUG nova.virt.hardware [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.267197] env[63372]: DEBUG nova.virt.hardware [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.267445] env[63372]: DEBUG nova.virt.hardware [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.268758] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a429ffe4-00ce-43c9-881a-fb0c888b3128 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.278094] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d61c822-e72e-4f8e-8938-0ed5506a8700 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.323535] env[63372]: DEBUG nova.scheduler.client.report [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.373283] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024055, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.451494] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024058, 'name': CreateVM_Task, 'duration_secs': 0.39991} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.452126] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 787.452924] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.458264] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520f3db2-5a55-ce1c-013e-1dc5197c9c5b, 'name': SearchDatastore_Task, 'duration_secs': 0.013922} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.458627] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.458845] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 787.459071] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.459313] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.459635] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 787.459919] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f4877df-5e34-4628-9d2f-c69dadde5937 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.466862] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 787.466862] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52a27815-657e-49dc-b6eb-78848cd8f050" [ 787.466862] env[63372]: _type = "Task" [ 787.466862] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.483942] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52aad9b5-248b-bc42-a84f-95a3ec9fb53d, 'name': SearchDatastore_Task, 'duration_secs': 0.015795} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.484410] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a27815-657e-49dc-b6eb-78848cd8f050, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.485275] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d92fe90-6d0e-4140-859f-5d6f7db5276d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.493191] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 787.493191] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52df1d76-b55b-1431-2f16-56a6e2c96a2f" [ 787.493191] env[63372]: _type = "Task" [ 787.493191] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.506039] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52df1d76-b55b-1431-2f16-56a6e2c96a2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.692261] env[63372]: DEBUG oslo_vmware.api [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1024059, 'name': PowerOffVM_Task, 'duration_secs': 0.315663} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.692586] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 787.692771] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 787.693049] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bde8ed42-4365-4e8e-943c-8a7142c8f648 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.698642] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "ac90a156-be00-4f62-a76e-e08914531167" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.698897] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "ac90a156-be00-4f62-a76e-e08914531167" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.764304] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 787.764520] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 787.764795] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Deleting the datastore file [datastore2] 7a75a51d-b804-453b-ba7b-fda6c4931802 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 787.765484] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-70bf9992-dd15-4397-baba-cf5bdc9aa1cc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.772906] env[63372]: DEBUG oslo_vmware.api [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Waiting for the task: (returnval){ [ 787.772906] env[63372]: value = "task-1024061" [ 787.772906] env[63372]: _type = "Task" [ 787.772906] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.783624] env[63372]: DEBUG oslo_vmware.api [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1024061, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.829175] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.632s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.829812] env[63372]: DEBUG nova.compute.manager [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 787.832616] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.260s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.832886] env[63372]: DEBUG nova.objects.instance [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lazy-loading 'resources' on Instance uuid 757df632-4af2-4eb3-bc6d-8c0812ba850d {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 787.872458] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024055, 'name': RemoveSnapshot_Task} progress is 98%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.902445] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "interface-bbba9f28-045d-41ab-8539-5b2968fe3d54-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.902822] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-bbba9f28-045d-41ab-8539-5b2968fe3d54-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.903255] env[63372]: DEBUG nova.objects.instance [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'flavor' on Instance uuid bbba9f28-045d-41ab-8539-5b2968fe3d54 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 787.935775] env[63372]: DEBUG nova.network.neutron [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Successfully updated port: 5bf209d9-9f64-410a-b14f-19cd792d4f7c {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 787.985617] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a27815-657e-49dc-b6eb-78848cd8f050, 'name': SearchDatastore_Task, 'duration_secs': 0.019516} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.986052] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.986341] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 787.986706] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.008254] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52df1d76-b55b-1431-2f16-56a6e2c96a2f, 'name': SearchDatastore_Task, 'duration_secs': 0.01756} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.008686] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.008974] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 3f66eda6-f5e9-4527-9711-849a01702580/3f66eda6-f5e9-4527-9711-849a01702580.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 788.009348] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.009593] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 788.009830] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b0e3d8c-1470-4bfa-9c02-765d30d2f254 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.012185] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-efdf0ddc-5e11-4e89-a6c5-04e41896fdb4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.022933] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 788.022933] env[63372]: value = "task-1024062" [ 788.022933] env[63372]: _type = "Task" [ 788.022933] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.024366] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 788.024592] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 788.028809] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c73534e0-a215-4c6d-a99d-c8cf2e38f17b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.037205] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024062, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.038697] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 788.038697] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5253a6b6-f6bf-2df9-0182-c244f8b863a1" [ 788.038697] env[63372]: _type = "Task" [ 788.038697] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.048984] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5253a6b6-f6bf-2df9-0182-c244f8b863a1, 'name': SearchDatastore_Task, 'duration_secs': 0.011202} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.049887] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc239946-42ab-4e35-929f-4b9eed70142c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.056519] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 788.056519] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]522c2edb-2947-18b4-bd49-cc7f44d1f049" [ 788.056519] env[63372]: _type = "Task" [ 788.056519] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.065409] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522c2edb-2947-18b4-bd49-cc7f44d1f049, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.091814] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "bad91d77-c7ee-4572-b1ed-068b2a55233c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.092044] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "bad91d77-c7ee-4572-b1ed-068b2a55233c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.092364] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "bad91d77-c7ee-4572-b1ed-068b2a55233c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.092610] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "bad91d77-c7ee-4572-b1ed-068b2a55233c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.094062] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "bad91d77-c7ee-4572-b1ed-068b2a55233c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.095301] env[63372]: INFO nova.compute.manager [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Terminating instance [ 788.097495] env[63372]: DEBUG nova.compute.manager [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 788.097764] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 788.098734] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9dab14-e0d5-4056-9a27-d29a57274a19 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.108163] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 788.108504] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa902b66-da62-486f-a62b-115fb3baca36 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.115857] env[63372]: DEBUG oslo_vmware.api [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 788.115857] env[63372]: value = "task-1024063" [ 788.115857] env[63372]: _type = "Task" [ 788.115857] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.125202] env[63372]: DEBUG oslo_vmware.api [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024063, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.202649] env[63372]: INFO nova.compute.manager [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Detaching volume e6717bc1-7823-4b07-889e-fd58898dee2a [ 788.243332] env[63372]: INFO nova.virt.block_device [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Attempting to driver detach volume e6717bc1-7823-4b07-889e-fd58898dee2a from mountpoint /dev/sdb [ 788.243607] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Volume detach. Driver type: vmdk {{(pid=63372) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 788.243858] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227348', 'volume_id': 'e6717bc1-7823-4b07-889e-fd58898dee2a', 'name': 'volume-e6717bc1-7823-4b07-889e-fd58898dee2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ac90a156-be00-4f62-a76e-e08914531167', 'attached_at': '', 'detached_at': '', 'volume_id': 'e6717bc1-7823-4b07-889e-fd58898dee2a', 'serial': 'e6717bc1-7823-4b07-889e-fd58898dee2a'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 788.244934] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fdd69c-9421-4914-8017-0982fc7f8d8d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.268418] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe465fe-a8a0-4ca4-8505-1d1c2a0c89e7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.280497] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89fb04da-fbfb-4925-9780-df8493b8b955 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.286560] env[63372]: DEBUG oslo_vmware.api [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Task: {'id': task-1024061, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199561} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.286961] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 788.287192] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 788.287405] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 788.287580] env[63372]: INFO nova.compute.manager [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Took 1.14 seconds to destroy the instance on the hypervisor. [ 788.287892] env[63372]: DEBUG oslo.service.loopingcall [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.288116] env[63372]: DEBUG nova.compute.manager [-] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 788.288218] env[63372]: DEBUG nova.network.neutron [-] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 788.307703] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71daf109-3047-4157-ae12-032dd8d4a762 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.328238] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] The volume has not been displaced from its original location: [datastore1] volume-e6717bc1-7823-4b07-889e-fd58898dee2a/volume-e6717bc1-7823-4b07-889e-fd58898dee2a.vmdk. No consolidation needed. {{(pid=63372) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 788.334381] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Reconfiguring VM instance instance-00000021 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 788.335955] env[63372]: DEBUG nova.compute.utils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 788.340397] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7161c6af-8a79-4fd7-a606-fc85dd73127d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.360940] env[63372]: DEBUG nova.compute.manager [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 788.361141] env[63372]: DEBUG nova.network.neutron [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 788.372760] env[63372]: DEBUG oslo_vmware.api [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 788.372760] env[63372]: value = "task-1024064" [ 788.372760] env[63372]: _type = "Task" [ 788.372760] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.377119] env[63372]: DEBUG oslo_vmware.api [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024055, 'name': RemoveSnapshot_Task, 'duration_secs': 1.172967} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.381028] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 788.381389] env[63372]: INFO nova.compute.manager [None req-3c3cfa0e-ba14-41cf-aec2-dde2e63b23c5 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Took 17.32 seconds to snapshot the instance on the hypervisor. [ 788.395371] env[63372]: DEBUG oslo_vmware.api [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024064, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.416358] env[63372]: DEBUG nova.policy [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d247cdc706e4f48a38c6ed1852b7a85', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1cc87c31eb77441daeec0fad5bb1ce87', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 788.439025] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "refresh_cache-de1b38a4-c7f2-420c-a050-7311976e4ca8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.439076] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquired lock "refresh_cache-de1b38a4-c7f2-420c-a050-7311976e4ca8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.439206] env[63372]: DEBUG nova.network.neutron [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 788.462355] env[63372]: DEBUG oslo_vmware.rw_handles [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b2d8ec-be36-d0c8-f291-0214a6c994dc/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 788.463507] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c947eb5-6eb1-470f-aebd-a0cce49f762c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.482495] env[63372]: DEBUG oslo_vmware.rw_handles [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b2d8ec-be36-d0c8-f291-0214a6c994dc/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 788.482714] env[63372]: ERROR oslo_vmware.rw_handles [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b2d8ec-be36-d0c8-f291-0214a6c994dc/disk-0.vmdk due to incomplete transfer. [ 788.485768] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-44a60508-055f-46ac-975b-e509a48b893b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.496651] env[63372]: DEBUG oslo_vmware.rw_handles [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b2d8ec-be36-d0c8-f291-0214a6c994dc/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 788.497199] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Uploaded image ec1d25a7-8ac1-4e60-bb42-a9bea2ae617a to the Glance image server {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 788.500049] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 788.508161] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ce99a5cc-a546-4e64-8ec9-f489a88e582f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.525867] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 788.525867] env[63372]: value = "task-1024065" [ 788.525867] env[63372]: _type = "Task" [ 788.525867] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.550762] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024062, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.556474] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024065, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.576458] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522c2edb-2947-18b4-bd49-cc7f44d1f049, 'name': SearchDatastore_Task, 'duration_secs': 0.017418} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.576921] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.577413] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] f6675697-1529-46be-b28a-398ff3060d18/f6675697-1529-46be-b28a-398ff3060d18.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 788.577878] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.578208] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 788.578569] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc6f7ce8-c1a1-4d1e-826f-e0a713b0a417 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.581891] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fafd68ed-4b8c-4345-ae59-fa6fb5d83394 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.596977] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 788.596977] env[63372]: value = "task-1024066" [ 788.596977] env[63372]: _type = "Task" [ 788.596977] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.608703] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 788.609029] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 788.610196] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30f22d60-3729-4a87-b53c-69bb860045e8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.619448] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024066, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.631354] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 788.631354] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]527962e5-3e6d-4c31-0beb-2c2dc6fe1aa8" [ 788.631354] env[63372]: _type = "Task" [ 788.631354] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.645717] env[63372]: DEBUG oslo_vmware.api [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024063, 'name': PowerOffVM_Task, 'duration_secs': 0.252311} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.650411] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 788.650769] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 788.651451] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-672d4471-c1f6-4093-b256-95c981525a9e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.659038] env[63372]: DEBUG oslo_concurrency.lockutils [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "cbd55518-a3b2-4636-ba43-c279fdd0bf8d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.659484] env[63372]: DEBUG oslo_concurrency.lockutils [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "cbd55518-a3b2-4636-ba43-c279fdd0bf8d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.659823] env[63372]: DEBUG oslo_concurrency.lockutils [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "cbd55518-a3b2-4636-ba43-c279fdd0bf8d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.660152] env[63372]: DEBUG oslo_concurrency.lockutils [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "cbd55518-a3b2-4636-ba43-c279fdd0bf8d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.660463] env[63372]: DEBUG oslo_concurrency.lockutils [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "cbd55518-a3b2-4636-ba43-c279fdd0bf8d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.662875] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527962e5-3e6d-4c31-0beb-2c2dc6fe1aa8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.663550] env[63372]: INFO nova.compute.manager [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Terminating instance [ 788.669962] env[63372]: DEBUG nova.objects.instance [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'pci_requests' on Instance uuid bbba9f28-045d-41ab-8539-5b2968fe3d54 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 788.675124] env[63372]: DEBUG nova.compute.manager [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 788.675124] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 788.675124] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-636329b4-d809-4083-b159-38cced043673 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.689219] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 788.689815] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3908951f-427c-4ba6-9207-805fc761c3fc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.698934] env[63372]: DEBUG oslo_vmware.api [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 788.698934] env[63372]: value = "task-1024068" [ 788.698934] env[63372]: _type = "Task" [ 788.698934] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.716721] env[63372]: DEBUG oslo_vmware.api [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024068, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.751242] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 788.751581] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 788.751915] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Deleting the datastore file [datastore2] bad91d77-c7ee-4572-b1ed-068b2a55233c {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 788.752236] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a30c31f9-f125-4b76-ac61-2ebf226c7c03 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.763728] env[63372]: DEBUG oslo_vmware.api [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 788.763728] env[63372]: value = "task-1024069" [ 788.763728] env[63372]: _type = "Task" [ 788.763728] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.776904] env[63372]: DEBUG oslo_vmware.api [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024069, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.813583] env[63372]: DEBUG nova.network.neutron [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Successfully created port: a0cd67cc-8bd5-4d24-8620-042e1fecd124 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.861697] env[63372]: DEBUG nova.compute.manager [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 788.900742] env[63372]: DEBUG oslo_vmware.api [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024064, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.012119] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06eccbf6-bab4-40ec-8f6b-7c46e2d4cb0c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.015629] env[63372]: DEBUG nova.network.neutron [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 789.022171] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fecc9d73-21de-4517-95b6-354186094a73 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.033769] env[63372]: DEBUG nova.compute.manager [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Received event network-vif-deleted-485f2c37-5aa1-470e-8b39-a986ebd7fd24 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.034026] env[63372]: DEBUG nova.compute.manager [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Received event network-vif-plugged-0e3060b6-6d21-4a2d-8e8a-ae48e64ee869 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.034226] env[63372]: DEBUG oslo_concurrency.lockutils [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] Acquiring lock "99f901a6-9bb3-4403-af0c-c8900f655cb3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.034431] env[63372]: DEBUG oslo_concurrency.lockutils [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] Lock "99f901a6-9bb3-4403-af0c-c8900f655cb3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.034598] env[63372]: DEBUG oslo_concurrency.lockutils [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] Lock "99f901a6-9bb3-4403-af0c-c8900f655cb3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.034761] env[63372]: DEBUG nova.compute.manager [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] No waiting events found dispatching network-vif-plugged-0e3060b6-6d21-4a2d-8e8a-ae48e64ee869 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 789.034938] env[63372]: WARNING nova.compute.manager [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Received unexpected event network-vif-plugged-0e3060b6-6d21-4a2d-8e8a-ae48e64ee869 for instance with vm_state building and task_state spawning. [ 789.035119] env[63372]: DEBUG nova.compute.manager [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Received event network-changed-0e3060b6-6d21-4a2d-8e8a-ae48e64ee869 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.035285] env[63372]: DEBUG nova.compute.manager [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Refreshing instance network info cache due to event network-changed-0e3060b6-6d21-4a2d-8e8a-ae48e64ee869. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 789.035473] env[63372]: DEBUG oslo_concurrency.lockutils [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] Acquiring lock "refresh_cache-99f901a6-9bb3-4403-af0c-c8900f655cb3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.035610] env[63372]: DEBUG oslo_concurrency.lockutils [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] Acquired lock "refresh_cache-99f901a6-9bb3-4403-af0c-c8900f655cb3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.035794] env[63372]: DEBUG nova.network.neutron [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Refreshing network info cache for port 0e3060b6-6d21-4a2d-8e8a-ae48e64ee869 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 789.071990] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c54c8e4-414b-49e5-92c5-77b936979d94 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.080716] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024062, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.690861} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.083169] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 3f66eda6-f5e9-4527-9711-849a01702580/3f66eda6-f5e9-4527-9711-849a01702580.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 789.083472] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 789.084097] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-efa44d56-5533-4674-8b45-980520cd3215 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.092171] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024065, 'name': Destroy_Task, 'duration_secs': 0.525658} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.093790] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a36484f-f7fe-456d-adf2-7723600aa465 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.097777] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Destroyed the VM [ 789.098028] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 789.099451] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-18390bfa-435c-4b82-a243-b97720026af4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.105443] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 789.105443] env[63372]: value = "task-1024070" [ 789.105443] env[63372]: _type = "Task" [ 789.105443] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.122457] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 789.122457] env[63372]: value = "task-1024071" [ 789.122457] env[63372]: _type = "Task" [ 789.122457] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.122699] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024066, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.123541] env[63372]: DEBUG nova.compute.provider_tree [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.137691] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024070, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.154377] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024071, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.158620] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527962e5-3e6d-4c31-0beb-2c2dc6fe1aa8, 'name': SearchDatastore_Task, 'duration_secs': 0.044754} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.159400] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da5f7f5e-e511-406a-b4e6-cb357c3f1ba1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.165462] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 789.165462] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52c4ded1-c815-76c7-2dea-83278370d5a7" [ 789.165462] env[63372]: _type = "Task" [ 789.165462] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.176262] env[63372]: DEBUG nova.objects.base [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 789.176500] env[63372]: DEBUG nova.network.neutron [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 789.178512] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c4ded1-c815-76c7-2dea-83278370d5a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.212230] env[63372]: DEBUG oslo_vmware.api [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024068, 'name': PowerOffVM_Task, 'duration_secs': 0.222849} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.213080] env[63372]: DEBUG nova.network.neutron [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Port 874f14b1-6cc3-4701-8a08-35940fc9bd9b binding to destination host cpu-1 is already ACTIVE {{(pid=63372) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 789.214038] env[63372]: DEBUG oslo_concurrency.lockutils [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.214038] env[63372]: DEBUG oslo_concurrency.lockutils [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquired lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.214038] env[63372]: DEBUG nova.network.neutron [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 789.214905] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 789.218176] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 789.218332] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8909b90c-a9d6-48b5-bdf2-67175bfa8161 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.278893] env[63372]: DEBUG oslo_vmware.api [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024069, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.301771] env[63372]: DEBUG nova.policy [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7ffe6776b664c00b9086046f9171ec6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46d2cb188a004dc385f2aa15458be929', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 789.337233] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 789.337526] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 789.337732] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Deleting the datastore file [datastore1] cbd55518-a3b2-4636-ba43-c279fdd0bf8d {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 789.338039] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95ebf245-b972-45ec-b20c-1f2721be2170 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.356421] env[63372]: DEBUG oslo_vmware.api [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 789.356421] env[63372]: value = "task-1024073" [ 789.356421] env[63372]: _type = "Task" [ 789.356421] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.377502] env[63372]: DEBUG oslo_vmware.api [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024073, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.393934] env[63372]: DEBUG oslo_vmware.api [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024064, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.470848] env[63372]: DEBUG nova.network.neutron [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Updating instance_info_cache with network_info: [{"id": "5bf209d9-9f64-410a-b14f-19cd792d4f7c", "address": "fa:16:3e:d0:73:0c", "network": {"id": "8171761d-bd68-4efd-ad4f-e661badde0c4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1048168720-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1cc87c31eb77441daeec0fad5bb1ce87", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bf209d9-9f", "ovs_interfaceid": "5bf209d9-9f64-410a-b14f-19cd792d4f7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.581322] env[63372]: DEBUG nova.network.neutron [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Updated VIF entry in instance network info cache for port 0e3060b6-6d21-4a2d-8e8a-ae48e64ee869. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 789.581692] env[63372]: DEBUG nova.network.neutron [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Updating instance_info_cache with network_info: [{"id": "0e3060b6-6d21-4a2d-8e8a-ae48e64ee869", "address": "fa:16:3e:81:de:69", "network": {"id": "8171761d-bd68-4efd-ad4f-e661badde0c4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1048168720-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1cc87c31eb77441daeec0fad5bb1ce87", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e3060b6-6d", "ovs_interfaceid": "0e3060b6-6d21-4a2d-8e8a-ae48e64ee869", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.584371] env[63372]: DEBUG nova.network.neutron [-] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.609582] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024066, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.947566} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.613019] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] f6675697-1529-46be-b28a-398ff3060d18/f6675697-1529-46be-b28a-398ff3060d18.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 789.613259] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 789.613594] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de2bb97e-9e38-44e4-8da4-2963c916e684 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.621693] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024070, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.34909} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.622815] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 789.623210] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 789.623210] env[63372]: value = "task-1024074" [ 789.623210] env[63372]: _type = "Task" [ 789.623210] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.623954] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e155ef9f-402b-4fea-b73a-9a933b988266 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.632017] env[63372]: DEBUG nova.scheduler.client.report [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.657036] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 3f66eda6-f5e9-4527-9711-849a01702580/3f66eda6-f5e9-4527-9711-849a01702580.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.661827] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.829s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.664101] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c45d6d1-b41d-412b-890c-ea6f3e205f40 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.678567] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024074, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.679062] env[63372]: DEBUG oslo_concurrency.lockutils [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.915s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.679142] env[63372]: DEBUG nova.objects.instance [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lazy-loading 'resources' on Instance uuid 7cf6ac9b-4703-4143-a22c-abc5528ce5ef {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 789.690232] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024071, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.692183] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 789.692183] env[63372]: value = "task-1024075" [ 789.692183] env[63372]: _type = "Task" [ 789.692183] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.700021] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c4ded1-c815-76c7-2dea-83278370d5a7, 'name': SearchDatastore_Task, 'duration_secs': 0.068298} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.700021] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.701223] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 99f901a6-9bb3-4403-af0c-c8900f655cb3/99f901a6-9bb3-4403-af0c-c8900f655cb3.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 789.701223] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54a7327d-37b9-445a-8c39-6824638abcc7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.709500] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024075, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.711227] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 789.711227] env[63372]: value = "task-1024076" [ 789.711227] env[63372]: _type = "Task" [ 789.711227] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.712297] env[63372]: INFO nova.scheduler.client.report [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleted allocations for instance 757df632-4af2-4eb3-bc6d-8c0812ba850d [ 789.728404] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024076, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.778908] env[63372]: DEBUG oslo_vmware.api [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024069, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.537023} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.778908] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 789.778908] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 789.778908] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 789.778908] env[63372]: INFO nova.compute.manager [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Took 1.68 seconds to destroy the instance on the hypervisor. [ 789.778908] env[63372]: DEBUG oslo.service.loopingcall [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.778908] env[63372]: DEBUG nova.compute.manager [-] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 789.778908] env[63372]: DEBUG nova.network.neutron [-] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 789.842597] env[63372]: DEBUG nova.network.neutron [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Successfully created port: cd1ce4d9-46e7-41b8-8009-71283aa762b2 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.874135] env[63372]: DEBUG nova.compute.manager [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 789.882501] env[63372]: DEBUG oslo_vmware.api [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024073, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.894448] env[63372]: DEBUG oslo_vmware.api [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024064, 'name': ReconfigVM_Task, 'duration_secs': 1.350835} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.894448] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Reconfigured VM instance instance-00000021 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 789.898844] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95863d5c-fb33-4529-8b04-337ce7b13986 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.912214] env[63372]: DEBUG nova.virt.hardware [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 789.912214] env[63372]: DEBUG nova.virt.hardware [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 789.912214] env[63372]: DEBUG nova.virt.hardware [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 789.912442] env[63372]: DEBUG nova.virt.hardware [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 789.912546] env[63372]: DEBUG nova.virt.hardware [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 789.912712] env[63372]: DEBUG nova.virt.hardware [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 789.912923] env[63372]: DEBUG nova.virt.hardware [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 789.913111] env[63372]: DEBUG nova.virt.hardware [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 789.913292] env[63372]: DEBUG nova.virt.hardware [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 789.913476] env[63372]: DEBUG nova.virt.hardware [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 789.913654] env[63372]: DEBUG nova.virt.hardware [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.921022] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698884c5-b629-4c39-a9a5-ab14a0c6bc57 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.930738] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-877a9420-baf2-49fe-9eea-7dadc5049556 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.934452] env[63372]: DEBUG oslo_vmware.api [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 789.934452] env[63372]: value = "task-1024077" [ 789.934452] env[63372]: _type = "Task" [ 789.934452] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.952261] env[63372]: DEBUG oslo_vmware.api [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024077, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.976675] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Releasing lock "refresh_cache-de1b38a4-c7f2-420c-a050-7311976e4ca8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.977940] env[63372]: DEBUG nova.compute.manager [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Instance network_info: |[{"id": "5bf209d9-9f64-410a-b14f-19cd792d4f7c", "address": "fa:16:3e:d0:73:0c", "network": {"id": "8171761d-bd68-4efd-ad4f-e661badde0c4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1048168720-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1cc87c31eb77441daeec0fad5bb1ce87", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bf209d9-9f", "ovs_interfaceid": "5bf209d9-9f64-410a-b14f-19cd792d4f7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 789.978484] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:73:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5bf209d9-9f64-410a-b14f-19cd792d4f7c', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 789.986933] env[63372]: DEBUG oslo.service.loopingcall [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.987942] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 789.988457] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-62a460b8-eb2b-42ae-b5f0-a0d18e1defac {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.010214] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 790.010214] env[63372]: value = "task-1024078" [ 790.010214] env[63372]: _type = "Task" [ 790.010214] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.017414] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024078, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.086506] env[63372]: DEBUG oslo_concurrency.lockutils [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] Releasing lock "refresh_cache-99f901a6-9bb3-4403-af0c-c8900f655cb3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.086506] env[63372]: DEBUG nova.compute.manager [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Received event network-vif-plugged-5bf209d9-9f64-410a-b14f-19cd792d4f7c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.086506] env[63372]: DEBUG oslo_concurrency.lockutils [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] Acquiring lock "de1b38a4-c7f2-420c-a050-7311976e4ca8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.086506] env[63372]: DEBUG oslo_concurrency.lockutils [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] Lock "de1b38a4-c7f2-420c-a050-7311976e4ca8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.086506] env[63372]: DEBUG oslo_concurrency.lockutils [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] Lock "de1b38a4-c7f2-420c-a050-7311976e4ca8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.086506] env[63372]: DEBUG nova.compute.manager [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] No waiting events found dispatching network-vif-plugged-5bf209d9-9f64-410a-b14f-19cd792d4f7c {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 790.086506] env[63372]: WARNING nova.compute.manager [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Received unexpected event network-vif-plugged-5bf209d9-9f64-410a-b14f-19cd792d4f7c for instance with vm_state building and task_state spawning. [ 790.086506] env[63372]: DEBUG nova.compute.manager [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Received event network-changed-5bf209d9-9f64-410a-b14f-19cd792d4f7c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.087179] env[63372]: DEBUG nova.compute.manager [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Refreshing instance network info cache due to event network-changed-5bf209d9-9f64-410a-b14f-19cd792d4f7c. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 790.087179] env[63372]: DEBUG oslo_concurrency.lockutils [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] Acquiring lock "refresh_cache-de1b38a4-c7f2-420c-a050-7311976e4ca8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.087179] env[63372]: DEBUG oslo_concurrency.lockutils [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] Acquired lock "refresh_cache-de1b38a4-c7f2-420c-a050-7311976e4ca8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.087179] env[63372]: DEBUG nova.network.neutron [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Refreshing network info cache for port 5bf209d9-9f64-410a-b14f-19cd792d4f7c {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 790.088451] env[63372]: INFO nova.compute.manager [-] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Took 1.80 seconds to deallocate network for instance. [ 790.149886] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024074, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090681} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.150377] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 790.150726] env[63372]: DEBUG oslo_vmware.api [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024071, 'name': RemoveSnapshot_Task, 'duration_secs': 0.639118} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.151732] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692fabe5-8903-4fc3-b057-54ec99f0c9fe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.155179] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 790.155179] env[63372]: INFO nova.compute.manager [None req-20ef41aa-c6b1-4fe9-8d65-d38380aa55ef tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Took 16.89 seconds to snapshot the instance on the hypervisor. [ 790.180246] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] f6675697-1529-46be-b28a-398ff3060d18/f6675697-1529-46be-b28a-398ff3060d18.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 790.184020] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-213f7c07-85d6-4686-a77d-6eab5b51b8a3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.209964] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024075, 'name': ReconfigVM_Task, 'duration_secs': 0.311953} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.211406] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 3f66eda6-f5e9-4527-9711-849a01702580/3f66eda6-f5e9-4527-9711-849a01702580.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 790.212110] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 790.212110] env[63372]: value = "task-1024079" [ 790.212110] env[63372]: _type = "Task" [ 790.212110] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.212308] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a085bece-f09e-4335-b89a-725ad05a0d8c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.235872] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024079, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.236349] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 790.236349] env[63372]: value = "task-1024080" [ 790.236349] env[63372]: _type = "Task" [ 790.236349] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.237029] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0250e264-b53c-4e07-8b9f-953943a9c753 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "757df632-4af2-4eb3-bc6d-8c0812ba850d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.162s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.239176] env[63372]: DEBUG nova.network.neutron [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance_info_cache with network_info: [{"id": "874f14b1-6cc3-4701-8a08-35940fc9bd9b", "address": "fa:16:3e:57:46:8e", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.24", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap874f14b1-6c", "ovs_interfaceid": "874f14b1-6cc3-4701-8a08-35940fc9bd9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.250845] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024076, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.253098] env[63372]: DEBUG oslo_concurrency.lockutils [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Releasing lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.264924] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024080, 'name': Rename_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.347502] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5edd972d-e251-4f7e-a052-5cc03ad3d916 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "93a5d948-0629-4f53-a681-858d519acfa7" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.347502] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5edd972d-e251-4f7e-a052-5cc03ad3d916 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "93a5d948-0629-4f53-a681-858d519acfa7" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.349026] env[63372]: DEBUG nova.compute.manager [None req-5edd972d-e251-4f7e-a052-5cc03ad3d916 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 790.349774] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3924a2d-288f-4445-84a5-8a107741f3be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.370239] env[63372]: DEBUG nova.compute.manager [None req-5edd972d-e251-4f7e-a052-5cc03ad3d916 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63372) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 790.370239] env[63372]: DEBUG nova.objects.instance [None req-5edd972d-e251-4f7e-a052-5cc03ad3d916 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lazy-loading 'flavor' on Instance uuid 93a5d948-0629-4f53-a681-858d519acfa7 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 790.378177] env[63372]: DEBUG oslo_vmware.api [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024073, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.56465} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.379121] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 790.379121] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 790.379286] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 790.379448] env[63372]: INFO nova.compute.manager [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Took 1.71 seconds to destroy the instance on the hypervisor. [ 790.379778] env[63372]: DEBUG oslo.service.loopingcall [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.383407] env[63372]: DEBUG nova.compute.manager [-] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 790.383525] env[63372]: DEBUG nova.network.neutron [-] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 790.454224] env[63372]: DEBUG oslo_vmware.api [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024077, 'name': ReconfigVM_Task, 'duration_secs': 0.158635} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.454879] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227348', 'volume_id': 'e6717bc1-7823-4b07-889e-fd58898dee2a', 'name': 'volume-e6717bc1-7823-4b07-889e-fd58898dee2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ac90a156-be00-4f62-a76e-e08914531167', 'attached_at': '', 'detached_at': '', 'volume_id': 'e6717bc1-7823-4b07-889e-fd58898dee2a', 'serial': 'e6717bc1-7823-4b07-889e-fd58898dee2a'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 790.520828] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024078, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.596789] env[63372]: DEBUG oslo_concurrency.lockutils [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.736643] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024076, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.92877} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.745496] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 99f901a6-9bb3-4403-af0c-c8900f655cb3/99f901a6-9bb3-4403-af0c-c8900f655cb3.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 790.745496] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 790.745496] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024079, 'name': ReconfigVM_Task, 'duration_secs': 0.510802} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.745496] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d2ffbd3-3c96-47bb-84e5-ff098d96939e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.752614] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Reconfigured VM instance instance-0000003d to attach disk [datastore1] f6675697-1529-46be-b28a-398ff3060d18/f6675697-1529-46be-b28a-398ff3060d18.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 790.752614] env[63372]: DEBUG nova.network.neutron [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Successfully updated port: a0cd67cc-8bd5-4d24-8620-042e1fecd124 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 790.756788] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7789c264-ab70-49f6-844e-25c5da9478bc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.759396] env[63372]: DEBUG nova.compute.manager [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63372) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 790.759724] env[63372]: DEBUG oslo_concurrency.lockutils [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.771877] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024080, 'name': Rename_Task, 'duration_secs': 0.22459} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.775124] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 790.775652] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 790.775652] env[63372]: value = "task-1024081" [ 790.775652] env[63372]: _type = "Task" [ 790.775652] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.776140] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 790.776140] env[63372]: value = "task-1024082" [ 790.776140] env[63372]: _type = "Task" [ 790.776140] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.776291] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3e88b2c9-8860-484a-ae90-2b53801a664a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.788154] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d5273c-f056-4296-b07e-e59543bbeed4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.808312] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3135eb12-c2ac-4177-8e3d-e0dd9a256c22 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.811894] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 790.811894] env[63372]: value = "task-1024083" [ 790.811894] env[63372]: _type = "Task" [ 790.811894] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.812164] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024082, 'name': Rename_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.812400] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024081, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.847252] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e59481a-d234-4109-8cae-4c0a4c36d522 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.853856] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024083, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.866533] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8269b20b-8dcd-45d4-8e50-5d2a9290c72d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.882331] env[63372]: DEBUG nova.compute.provider_tree [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 184, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 790.886792] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5edd972d-e251-4f7e-a052-5cc03ad3d916 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 790.886792] env[63372]: DEBUG nova.network.neutron [-] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.886792] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79828798-d8c7-43de-b332-c4b20b25b99f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.894015] env[63372]: DEBUG oslo_vmware.api [None req-5edd972d-e251-4f7e-a052-5cc03ad3d916 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 790.894015] env[63372]: value = "task-1024084" [ 790.894015] env[63372]: _type = "Task" [ 790.894015] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.901308] env[63372]: DEBUG nova.network.neutron [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Updated VIF entry in instance network info cache for port 5bf209d9-9f64-410a-b14f-19cd792d4f7c. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 790.901308] env[63372]: DEBUG nova.network.neutron [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Updating instance_info_cache with network_info: [{"id": "5bf209d9-9f64-410a-b14f-19cd792d4f7c", "address": "fa:16:3e:d0:73:0c", "network": {"id": "8171761d-bd68-4efd-ad4f-e661badde0c4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1048168720-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1cc87c31eb77441daeec0fad5bb1ce87", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bf209d9-9f", "ovs_interfaceid": "5bf209d9-9f64-410a-b14f-19cd792d4f7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.908171] env[63372]: DEBUG oslo_vmware.api [None req-5edd972d-e251-4f7e-a052-5cc03ad3d916 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024084, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.021439] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024078, 'name': CreateVM_Task, 'duration_secs': 0.869547} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.021661] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 791.022499] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.022690] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.023074] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 791.023370] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa0ea595-2df5-4c45-955c-9d0008d4634f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.029555] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 791.029555] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52a935d7-0350-e767-293d-42cd53a63d26" [ 791.029555] env[63372]: _type = "Task" [ 791.029555] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.039132] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a935d7-0350-e767-293d-42cd53a63d26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.044046] env[63372]: DEBUG nova.objects.instance [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lazy-loading 'flavor' on Instance uuid ac90a156-be00-4f62-a76e-e08914531167 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 791.205365] env[63372]: DEBUG nova.network.neutron [-] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.257763] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "refresh_cache-7c9d551e-d210-4943-971f-b2829751fcfc" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.259594] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquired lock "refresh_cache-7c9d551e-d210-4943-971f-b2829751fcfc" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.259594] env[63372]: DEBUG nova.network.neutron [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.292216] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024082, 'name': Rename_Task, 'duration_secs': 0.15754} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.295075] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 791.295388] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024081, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072781} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.295857] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15cd1522-953a-4d06-b43c-de7ad6f16446 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.297505] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 791.298263] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ffd574-a309-4a82-a1be-f26f687a5a4f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.321849] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 99f901a6-9bb3-4403-af0c-c8900f655cb3/99f901a6-9bb3-4403-af0c-c8900f655cb3.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 791.326226] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73016532-edb1-4d6f-a31e-3b95e5103679 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.340716] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 791.340716] env[63372]: value = "task-1024085" [ 791.340716] env[63372]: _type = "Task" [ 791.340716] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.348474] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024083, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.349631] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 791.349631] env[63372]: value = "task-1024086" [ 791.349631] env[63372]: _type = "Task" [ 791.349631] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.356172] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024085, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.361553] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024086, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.390982] env[63372]: INFO nova.compute.manager [-] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Took 1.61 seconds to deallocate network for instance. [ 791.406649] env[63372]: DEBUG oslo_concurrency.lockutils [req-be0f5d1b-b482-4e08-8887-90b5c32cbf0a req-71293536-586b-4475-8991-c374ffa72aa4 service nova] Releasing lock "refresh_cache-de1b38a4-c7f2-420c-a050-7311976e4ca8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.407480] env[63372]: DEBUG oslo_vmware.api [None req-5edd972d-e251-4f7e-a052-5cc03ad3d916 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024084, 'name': PowerOffVM_Task, 'duration_secs': 0.272864} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.408564] env[63372]: ERROR nova.scheduler.client.report [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [req-3c6f2529-3400-48bd-ba89-da4e46658526] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 184, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3c6f2529-3400-48bd-ba89-da4e46658526"}]} [ 791.408935] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5edd972d-e251-4f7e-a052-5cc03ad3d916 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 791.409131] env[63372]: DEBUG nova.compute.manager [None req-5edd972d-e251-4f7e-a052-5cc03ad3d916 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 791.412644] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7090de8-33cf-4d1e-b048-ddf7aa51402a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.430950] env[63372]: DEBUG nova.scheduler.client.report [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Refreshing inventories for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 791.445284] env[63372]: DEBUG nova.scheduler.client.report [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Updating ProviderTree inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 791.445564] env[63372]: DEBUG nova.compute.provider_tree [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 791.457480] env[63372]: DEBUG nova.scheduler.client.report [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Refreshing aggregate associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, aggregates: None {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 791.474784] env[63372]: DEBUG nova.scheduler.client.report [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Refreshing trait associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 791.540707] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a935d7-0350-e767-293d-42cd53a63d26, 'name': SearchDatastore_Task, 'duration_secs': 0.018951} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.547021] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.547021] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 791.547021] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.547021] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.547021] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 791.547021] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9cdf651e-7d4c-490d-ac88-23971aa6a508 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.563395] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 791.563505] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 791.565034] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78944298-eb83-48b4-9532-159530df9e6f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.572681] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 791.572681] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5271fe3f-9dd2-292f-8d13-93a16fc36075" [ 791.572681] env[63372]: _type = "Task" [ 791.572681] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.580623] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5271fe3f-9dd2-292f-8d13-93a16fc36075, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.668291] env[63372]: INFO nova.compute.manager [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Rebuilding instance [ 791.694915] env[63372]: DEBUG nova.network.neutron [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Successfully updated port: cd1ce4d9-46e7-41b8-8009-71283aa762b2 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 791.707680] env[63372]: INFO nova.compute.manager [-] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Took 1.32 seconds to deallocate network for instance. [ 791.719900] env[63372]: DEBUG nova.compute.manager [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 791.720789] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3c472d-4426-4529-a295-540a859734ea {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.793311] env[63372]: DEBUG nova.network.neutron [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 791.803890] env[63372]: DEBUG nova.compute.manager [req-b4c68e38-af8f-4757-9163-264b601fab56 req-b2731b0e-92c8-4512-8930-6b0c5a41de24 service nova] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Received event network-vif-plugged-a0cd67cc-8bd5-4d24-8620-042e1fecd124 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.803890] env[63372]: DEBUG oslo_concurrency.lockutils [req-b4c68e38-af8f-4757-9163-264b601fab56 req-b2731b0e-92c8-4512-8930-6b0c5a41de24 service nova] Acquiring lock "7c9d551e-d210-4943-971f-b2829751fcfc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.803890] env[63372]: DEBUG oslo_concurrency.lockutils [req-b4c68e38-af8f-4757-9163-264b601fab56 req-b2731b0e-92c8-4512-8930-6b0c5a41de24 service nova] Lock "7c9d551e-d210-4943-971f-b2829751fcfc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.803890] env[63372]: DEBUG oslo_concurrency.lockutils [req-b4c68e38-af8f-4757-9163-264b601fab56 req-b2731b0e-92c8-4512-8930-6b0c5a41de24 service nova] Lock "7c9d551e-d210-4943-971f-b2829751fcfc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.803890] env[63372]: DEBUG nova.compute.manager [req-b4c68e38-af8f-4757-9163-264b601fab56 req-b2731b0e-92c8-4512-8930-6b0c5a41de24 service nova] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] No waiting events found dispatching network-vif-plugged-a0cd67cc-8bd5-4d24-8620-042e1fecd124 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 791.803890] env[63372]: WARNING nova.compute.manager [req-b4c68e38-af8f-4757-9163-264b601fab56 req-b2731b0e-92c8-4512-8930-6b0c5a41de24 service nova] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Received unexpected event network-vif-plugged-a0cd67cc-8bd5-4d24-8620-042e1fecd124 for instance with vm_state building and task_state spawning. [ 791.823649] env[63372]: DEBUG oslo_vmware.api [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024083, 'name': PowerOnVM_Task, 'duration_secs': 0.633557} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.824180] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 791.824845] env[63372]: INFO nova.compute.manager [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Took 12.66 seconds to spawn the instance on the hypervisor. [ 791.826761] env[63372]: DEBUG nova.compute.manager [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 791.826761] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19888182-97e0-4e8f-be16-ec92ea80a938 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.854965] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024085, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.866795] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024086, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.901499] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.926402] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5edd972d-e251-4f7e-a052-5cc03ad3d916 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "93a5d948-0629-4f53-a681-858d519acfa7" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.579s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.930236] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97fbb2f9-497e-4ff5-9f9a-4e2126e04d35 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.941080] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7f244d-226b-4620-9b5e-160c15ba2859 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.973657] env[63372]: DEBUG nova.network.neutron [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Updating instance_info_cache with network_info: [{"id": "a0cd67cc-8bd5-4d24-8620-042e1fecd124", "address": "fa:16:3e:bb:a7:ca", "network": {"id": "8171761d-bd68-4efd-ad4f-e661badde0c4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1048168720-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1cc87c31eb77441daeec0fad5bb1ce87", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0cd67cc-8b", "ovs_interfaceid": "a0cd67cc-8bd5-4d24-8620-042e1fecd124", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.975980] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029b234e-7000-406c-90d6-efcc6441623c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.986209] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562f0f45-350d-4610-9a54-e6bb3d79c80c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.002063] env[63372]: DEBUG nova.compute.provider_tree [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 184, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 792.052493] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0ccbcfae-950f-47dd-97f7-5b006d0395cc tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "ac90a156-be00-4f62-a76e-e08914531167" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.353s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.083744] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5271fe3f-9dd2-292f-8d13-93a16fc36075, 'name': SearchDatastore_Task, 'duration_secs': 0.051344} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.084576] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40b3a10b-dc5a-4775-ac2e-48b2b872df9f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.090907] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 792.090907] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52abb50c-a7f1-84a8-856b-9bf79e0f4a68" [ 792.090907] env[63372]: _type = "Task" [ 792.090907] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.099431] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52abb50c-a7f1-84a8-856b-9bf79e0f4a68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.204016] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.204451] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.204727] env[63372]: DEBUG nova.network.neutron [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 792.216842] env[63372]: DEBUG oslo_concurrency.lockutils [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.239979] env[63372]: DEBUG nova.compute.manager [req-99f19004-9ca7-4b3b-8237-3ac6a937ed60 req-c6f43ae2-3ea2-4b16-8843-71a72275a636 service nova] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Received event network-vif-deleted-6c9c3162-8734-4911-a9d1-9f3f9214ea93 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.240153] env[63372]: DEBUG nova.compute.manager [req-99f19004-9ca7-4b3b-8237-3ac6a937ed60 req-c6f43ae2-3ea2-4b16-8843-71a72275a636 service nova] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Received event network-vif-deleted-49b9f908-9366-4318-b2d3-c408e404a9df {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.240387] env[63372]: DEBUG nova.compute.manager [req-99f19004-9ca7-4b3b-8237-3ac6a937ed60 req-c6f43ae2-3ea2-4b16-8843-71a72275a636 service nova] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Received event network-vif-deleted-116aaf8a-12e9-4e54-befb-6e8bc6d0a1ee {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.242444] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 792.242811] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a0ccf210-5f81-4890-acfe-760fc18f51b5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.255249] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Waiting for the task: (returnval){ [ 792.255249] env[63372]: value = "task-1024087" [ 792.255249] env[63372]: _type = "Task" [ 792.255249] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.265602] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1024087, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.358111] env[63372]: DEBUG oslo_vmware.api [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024085, 'name': PowerOnVM_Task, 'duration_secs': 1.019311} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.358737] env[63372]: INFO nova.compute.manager [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Took 43.53 seconds to build instance. [ 792.363778] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 792.364024] env[63372]: INFO nova.compute.manager [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Took 10.47 seconds to spawn the instance on the hypervisor. [ 792.364280] env[63372]: DEBUG nova.compute.manager [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 792.365312] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71b8da3-232e-496f-b00b-74534434193e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.374862] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024086, 'name': ReconfigVM_Task, 'duration_secs': 0.906557} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.376778] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 99f901a6-9bb3-4403-af0c-c8900f655cb3/99f901a6-9bb3-4403-af0c-c8900f655cb3.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 792.383036] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b9252a3e-fcba-437a-8041-0d4aa37b32b0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.389836] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 792.389836] env[63372]: value = "task-1024088" [ 792.389836] env[63372]: _type = "Task" [ 792.389836] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.399860] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024088, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.479297] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Releasing lock "refresh_cache-7c9d551e-d210-4943-971f-b2829751fcfc" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.480034] env[63372]: DEBUG nova.compute.manager [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Instance network_info: |[{"id": "a0cd67cc-8bd5-4d24-8620-042e1fecd124", "address": "fa:16:3e:bb:a7:ca", "network": {"id": "8171761d-bd68-4efd-ad4f-e661badde0c4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1048168720-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1cc87c31eb77441daeec0fad5bb1ce87", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0cd67cc-8b", "ovs_interfaceid": "a0cd67cc-8bd5-4d24-8620-042e1fecd124", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 792.480268] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:a7:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0cd67cc-8bd5-4d24-8620-042e1fecd124', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.490139] env[63372]: DEBUG oslo.service.loopingcall [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.490977] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 792.491352] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58bcfbcd-3ba1-4453-a846-e660040ed088 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.519548] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.519548] env[63372]: value = "task-1024089" [ 792.519548] env[63372]: _type = "Task" [ 792.519548] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.528459] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024089, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.541732] env[63372]: DEBUG nova.scheduler.client.report [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 84 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 184, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 792.541732] env[63372]: DEBUG nova.compute.provider_tree [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 84 to 85 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 792.541732] env[63372]: DEBUG nova.compute.provider_tree [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 184, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 792.604259] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52abb50c-a7f1-84a8-856b-9bf79e0f4a68, 'name': SearchDatastore_Task, 'duration_secs': 0.024152} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.604553] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.604817] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] de1b38a4-c7f2-420c-a050-7311976e4ca8/de1b38a4-c7f2-420c-a050-7311976e4ca8.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 792.605105] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0b9af86f-317b-47cb-bd4e-464db29da600 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.614247] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 792.614247] env[63372]: value = "task-1024090" [ 792.614247] env[63372]: _type = "Task" [ 792.614247] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.623296] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024090, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.767470] env[63372]: WARNING nova.network.neutron [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] e1f7ca1b-541e-468f-a439-a6841fc781ba already exists in list: networks containing: ['e1f7ca1b-541e-468f-a439-a6841fc781ba']. ignoring it [ 792.783517] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1024087, 'name': PowerOffVM_Task, 'duration_secs': 0.295667} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.783735] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 792.784510] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 792.784796] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e7b0612-1fca-47b0-9a78-4073f080e298 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.794862] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Waiting for the task: (returnval){ [ 792.794862] env[63372]: value = "task-1024091" [ 792.794862] env[63372]: _type = "Task" [ 792.794862] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.809237] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] VM already powered off {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 792.809769] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Volume detach. Driver type: vmdk {{(pid=63372) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 792.810519] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227234', 'volume_id': 'a483fa3d-7a97-48fa-8fe5-723336b546a9', 'name': 'volume-a483fa3d-7a97-48fa-8fe5-723336b546a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5d819bd2-f7f9-480d-83ae-fbd4d2e0d382', 'attached_at': '', 'detached_at': '', 'volume_id': 'a483fa3d-7a97-48fa-8fe5-723336b546a9', 'serial': 'a483fa3d-7a97-48fa-8fe5-723336b546a9'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 792.811786] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c9b1d2a-0f19-4e33-bd00-1a6aa3ebe350 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.842447] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41cc192-1afd-4f9f-aa54-8d83ea3ccb55 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.853467] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323d66e9-9cda-4aa5-aa27-7a5815e89247 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.875252] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2aa9f380-0bad-485b-ba7d-161a9e14e670 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "3f66eda6-f5e9-4527-9711-849a01702580" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.263s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.879126] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abeeb875-3aac-4eb4-bed2-2cfed36094ae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.900407] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] The volume has not been displaced from its original location: [datastore1] volume-a483fa3d-7a97-48fa-8fe5-723336b546a9/volume-a483fa3d-7a97-48fa-8fe5-723336b546a9.vmdk. No consolidation needed. {{(pid=63372) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 792.905945] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Reconfiguring VM instance instance-0000002b to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 792.906705] env[63372]: INFO nova.compute.manager [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Took 40.15 seconds to build instance. [ 792.910933] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f4313d9-3473-43b8-b704-fbf0e6e2f0cc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.928083] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3a6e326f-35cc-4629-a54f-3349da2f6d4d tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "f6675697-1529-46be-b28a-398ff3060d18" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.146s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.936852] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024088, 'name': Rename_Task, 'duration_secs': 0.21353} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.939018] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 792.940030] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Waiting for the task: (returnval){ [ 792.940030] env[63372]: value = "task-1024092" [ 792.940030] env[63372]: _type = "Task" [ 792.940030] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.940213] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d34f452-cbda-4f31-bc5e-2745ad711969 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.955426] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1024092, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.957712] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 792.957712] env[63372]: value = "task-1024093" [ 792.957712] env[63372]: _type = "Task" [ 792.957712] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.970115] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024093, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.036186] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024089, 'name': CreateVM_Task, 'duration_secs': 0.407979} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.036393] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 793.037286] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.037435] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.037799] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 793.038116] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3cf9956-aa11-4e7c-8788-4c6cbf69b253 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.044722] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 793.044722] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52442464-00c3-7b3f-5c0f-73b3eb6022c4" [ 793.044722] env[63372]: _type = "Task" [ 793.044722] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.053933] env[63372]: DEBUG oslo_concurrency.lockutils [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.375s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.056290] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.697s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.056471] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.056685] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63372) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 793.057126] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.809s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.059328] env[63372]: INFO nova.compute.claims [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.064453] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1128f3e-f00a-44af-b755-00463175acac {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.072986] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52442464-00c3-7b3f-5c0f-73b3eb6022c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.081655] env[63372]: INFO nova.scheduler.client.report [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Deleted allocations for instance 7cf6ac9b-4703-4143-a22c-abc5528ce5ef [ 793.086504] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60ba11b-786b-4cb0-952d-ccc54d774c0f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.106846] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13e8e26-f979-45e2-b6e4-0fd43a4cb108 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.121382] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2e404c-a0e2-4f43-977d-dc318566cbe8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.132535] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024090, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.165199] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178801MB free_disk=184GB free_vcpus=48 pci_devices=None {{(pid=63372) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 793.165529] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.238137] env[63372]: DEBUG nova.network.neutron [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Updating instance_info_cache with network_info: [{"id": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "address": "fa:16:3e:17:c8:2e", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4cf27a3-fc", "ovs_interfaceid": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cd1ce4d9-46e7-41b8-8009-71283aa762b2", "address": "fa:16:3e:32:b0:d7", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd1ce4d9-46", "ovs_interfaceid": "cd1ce4d9-46e7-41b8-8009-71283aa762b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.439680] env[63372]: DEBUG nova.compute.manager [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 793.440840] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-553d8026-abb9-4ea8-8d2a-9cb539e1d532 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.454368] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1024092, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.469294] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024093, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.558520] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52442464-00c3-7b3f-5c0f-73b3eb6022c4, 'name': SearchDatastore_Task, 'duration_secs': 0.058229} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.558647] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.558787] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.559036] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.559194] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.559377] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.559633] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1c128114-7777-4c74-bf20-73e75b031c78 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.568572] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.568758] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 793.569536] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24da5316-c6ea-4c4f-b7a0-d1d9967b9619 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.575586] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 793.575586] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5209ea4d-b919-deac-918d-a2f651f8a130" [ 793.575586] env[63372]: _type = "Task" [ 793.575586] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.584616] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5209ea4d-b919-deac-918d-a2f651f8a130, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.596849] env[63372]: DEBUG oslo_concurrency.lockutils [None req-516d49d7-73be-4add-8fe7-cb41d6d11d0d tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "7cf6ac9b-4703-4143-a22c-abc5528ce5ef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.397s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.627037] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024090, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542365} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.627417] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] de1b38a4-c7f2-420c-a050-7311976e4ca8/de1b38a4-c7f2-420c-a050-7311976e4ca8.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 793.627654] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 793.627906] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae7cc3ac-b8bf-4c36-8cf0-61b1ec2b7421 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.635419] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 793.635419] env[63372]: value = "task-1024094" [ 793.635419] env[63372]: _type = "Task" [ 793.635419] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.644085] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024094, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.742774] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.743547] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.743774] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.744974] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-615601e2-9eca-4c3f-864f-5438b47eb507 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.762708] env[63372]: DEBUG nova.virt.hardware [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 793.762956] env[63372]: DEBUG nova.virt.hardware [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 793.763125] env[63372]: DEBUG nova.virt.hardware [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 793.763310] env[63372]: DEBUG nova.virt.hardware [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 793.763462] env[63372]: DEBUG nova.virt.hardware [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 793.763614] env[63372]: DEBUG nova.virt.hardware [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 793.763820] env[63372]: DEBUG nova.virt.hardware [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 793.763982] env[63372]: DEBUG nova.virt.hardware [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 793.764167] env[63372]: DEBUG nova.virt.hardware [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 793.764336] env[63372]: DEBUG nova.virt.hardware [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 793.764511] env[63372]: DEBUG nova.virt.hardware [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 793.771057] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Reconfiguring VM to attach interface {{(pid=63372) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 793.771269] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-502c0986-dc87-407c-8b59-7bb953e4cfb9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.790846] env[63372]: DEBUG oslo_vmware.api [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 793.790846] env[63372]: value = "task-1024095" [ 793.790846] env[63372]: _type = "Task" [ 793.790846] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.799602] env[63372]: DEBUG oslo_vmware.api [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024095, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.956697] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1024092, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.958780] env[63372]: INFO nova.compute.manager [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] instance snapshotting [ 793.959439] env[63372]: DEBUG nova.objects.instance [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'flavor' on Instance uuid 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 793.971468] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024093, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.049041] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "4028d7c0-f398-4e43-9cff-5d89a14c4efa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.049248] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "4028d7c0-f398-4e43-9cff-5d89a14c4efa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.087915] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5209ea4d-b919-deac-918d-a2f651f8a130, 'name': SearchDatastore_Task, 'duration_secs': 0.008832} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.088794] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ccf7ce7-ad8a-4d44-bef6-89caa8de8631 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.099371] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 794.099371] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e97886-d36d-4688-fc1b-d0101fba9c25" [ 794.099371] env[63372]: _type = "Task" [ 794.099371] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.116355] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e97886-d36d-4688-fc1b-d0101fba9c25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.146989] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024094, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.310415] env[63372]: DEBUG oslo_vmware.api [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024095, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.334910] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "a3d5b93e-1b76-48e7-bb3e-2423e7c4119a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.335323] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "a3d5b93e-1b76-48e7-bb3e-2423e7c4119a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.366282] env[63372]: DEBUG oslo_concurrency.lockutils [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Acquiring lock "f43555ef-b517-4b7d-9d2e-4787d40e201b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.366648] env[63372]: DEBUG oslo_concurrency.lockutils [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Lock "f43555ef-b517-4b7d-9d2e-4787d40e201b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.366882] env[63372]: DEBUG oslo_concurrency.lockutils [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Acquiring lock "f43555ef-b517-4b7d-9d2e-4787d40e201b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.367096] env[63372]: DEBUG oslo_concurrency.lockutils [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Lock "f43555ef-b517-4b7d-9d2e-4787d40e201b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.367283] env[63372]: DEBUG oslo_concurrency.lockutils [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Lock "f43555ef-b517-4b7d-9d2e-4787d40e201b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.369905] env[63372]: INFO nova.compute.manager [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Terminating instance [ 794.371782] env[63372]: DEBUG nova.compute.manager [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 794.371982] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 794.372915] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8683ad47-5240-48a5-80d2-7d9577101650 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.381713] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 794.385162] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4b18428-145a-455b-9a9f-308518433a73 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.392658] env[63372]: DEBUG oslo_vmware.api [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 794.392658] env[63372]: value = "task-1024096" [ 794.392658] env[63372]: _type = "Task" [ 794.392658] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.405154] env[63372]: DEBUG oslo_vmware.api [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024096, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.457592] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1024092, 'name': ReconfigVM_Task, 'duration_secs': 1.329003} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.457901] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Reconfigured VM instance instance-0000002b to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 794.465342] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81428071-3fdc-4078-8d66-3c55ba3ebaeb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.483767] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32a4252-161a-4595-8791-2de3c0640fd4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.494055] env[63372]: DEBUG oslo_vmware.api [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024093, 'name': PowerOnVM_Task, 'duration_secs': 1.439754} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.494392] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Waiting for the task: (returnval){ [ 794.494392] env[63372]: value = "task-1024097" [ 794.494392] env[63372]: _type = "Task" [ 794.494392] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.507877] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 794.508155] env[63372]: INFO nova.compute.manager [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Took 9.94 seconds to spawn the instance on the hypervisor. [ 794.508351] env[63372]: DEBUG nova.compute.manager [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 794.512138] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3c6a6e5-020d-4d6d-aadd-85e2ed6dd719 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.518477] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c1b6cc-f26d-4016-ad04-a1fa2680bce3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.521961] env[63372]: DEBUG nova.objects.instance [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lazy-loading 'flavor' on Instance uuid 93a5d948-0629-4f53-a681-858d519acfa7 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 794.530424] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1024097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.553028] env[63372]: DEBUG nova.compute.manager [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.599275] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a657e67d-ba9f-40b2-b39a-0b9f4a568d92 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.619925] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16cfb8e6-b601-4507-b526-1f87efd19d83 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.623461] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e97886-d36d-4688-fc1b-d0101fba9c25, 'name': SearchDatastore_Task, 'duration_secs': 0.020982} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.623729] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.623991] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 7c9d551e-d210-4943-971f-b2829751fcfc/7c9d551e-d210-4943-971f-b2829751fcfc.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 794.624599] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b7682f8c-e855-452c-8623-d5645c56c7b9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.658275] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcfc28cc-0772-43ab-b1ed-684d656d814f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.661054] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 794.661054] env[63372]: value = "task-1024098" [ 794.661054] env[63372]: _type = "Task" [ 794.661054] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.672248] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024094, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.597771} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.672984] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 794.674215] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a670be8-f22c-4644-bff9-d448494abb85 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.682222] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a8c46c-8ba6-4eb0-91ad-0f0987d12de7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.684954] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024098, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.697121] env[63372]: DEBUG nova.compute.provider_tree [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.720024] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] de1b38a4-c7f2-420c-a050-7311976e4ca8/de1b38a4-c7f2-420c-a050-7311976e4ca8.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 794.720024] env[63372]: DEBUG nova.scheduler.client.report [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 184, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 794.723036] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24295b92-d3ee-4dff-856b-382830acaf66 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.744848] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 794.744848] env[63372]: value = "task-1024099" [ 794.744848] env[63372]: _type = "Task" [ 794.744848] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.755642] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024099, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.802739] env[63372]: DEBUG oslo_vmware.api [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024095, 'name': ReconfigVM_Task, 'duration_secs': 0.95671} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.804568] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.804873] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Reconfigured VM to attach interface {{(pid=63372) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 794.808701] env[63372]: DEBUG nova.compute.manager [req-8e5d9b1a-f840-4b04-871c-8a351e8db697 req-0c9e1588-5298-4e4f-b325-fcb5b0917063 service nova] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Received event network-changed-a0cd67cc-8bd5-4d24-8620-042e1fecd124 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 794.808924] env[63372]: DEBUG nova.compute.manager [req-8e5d9b1a-f840-4b04-871c-8a351e8db697 req-0c9e1588-5298-4e4f-b325-fcb5b0917063 service nova] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Refreshing instance network info cache due to event network-changed-a0cd67cc-8bd5-4d24-8620-042e1fecd124. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 794.809164] env[63372]: DEBUG oslo_concurrency.lockutils [req-8e5d9b1a-f840-4b04-871c-8a351e8db697 req-0c9e1588-5298-4e4f-b325-fcb5b0917063 service nova] Acquiring lock "refresh_cache-7c9d551e-d210-4943-971f-b2829751fcfc" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.809343] env[63372]: DEBUG oslo_concurrency.lockutils [req-8e5d9b1a-f840-4b04-871c-8a351e8db697 req-0c9e1588-5298-4e4f-b325-fcb5b0917063 service nova] Acquired lock "refresh_cache-7c9d551e-d210-4943-971f-b2829751fcfc" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.809522] env[63372]: DEBUG nova.network.neutron [req-8e5d9b1a-f840-4b04-871c-8a351e8db697 req-0c9e1588-5298-4e4f-b325-fcb5b0917063 service nova] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Refreshing network info cache for port a0cd67cc-8bd5-4d24-8620-042e1fecd124 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 794.838125] env[63372]: DEBUG nova.compute.manager [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.904333] env[63372]: DEBUG oslo_vmware.api [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024096, 'name': PowerOffVM_Task, 'duration_secs': 0.28934} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.904628] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 794.904804] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 794.905088] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-73d33da9-b534-4ec3-9766-ba7f656df347 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.988371] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 794.988575] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 794.990193] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Deleting the datastore file [datastore2] f43555ef-b517-4b7d-9d2e-4787d40e201b {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 794.990193] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bc12c0da-d863-421a-8e47-a0aab121f3f7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.996999] env[63372]: DEBUG oslo_vmware.api [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for the task: (returnval){ [ 794.996999] env[63372]: value = "task-1024101" [ 794.996999] env[63372]: _type = "Task" [ 794.996999] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.007180] env[63372]: DEBUG oslo_vmware.api [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024101, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.020098] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1024097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.028363] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.028563] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.028754] env[63372]: DEBUG nova.network.neutron [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 795.028943] env[63372]: DEBUG nova.objects.instance [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lazy-loading 'info_cache' on Instance uuid 93a5d948-0629-4f53-a681-858d519acfa7 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 795.047326] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 795.048878] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-44e656ae-1e50-430e-b611-a4f70e217416 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.054461] env[63372]: INFO nova.compute.manager [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Took 40.60 seconds to build instance. [ 795.066021] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 795.066021] env[63372]: value = "task-1024102" [ 795.066021] env[63372]: _type = "Task" [ 795.066021] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.078564] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024102, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.079748] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.173518] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024098, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.240731] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.183s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.241426] env[63372]: DEBUG nova.compute.manager [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 795.244386] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.582s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.246228] env[63372]: INFO nova.compute.claims [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 795.260563] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024099, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.311921] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0ca5c901-7c04-4fec-a748-e90a7e31b31d tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-bbba9f28-045d-41ab-8539-5b2968fe3d54-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.409s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.363618] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.510841] env[63372]: DEBUG oslo_vmware.api [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Task: {'id': task-1024101, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274501} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.515459] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 795.515773] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 795.516055] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 795.516324] env[63372]: INFO nova.compute.manager [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 795.516689] env[63372]: DEBUG oslo.service.loopingcall [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 795.516988] env[63372]: DEBUG nova.compute.manager [-] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 795.517138] env[63372]: DEBUG nova.network.neutron [-] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 795.529799] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1024097, 'name': ReconfigVM_Task, 'duration_secs': 0.906943} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.530174] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227234', 'volume_id': 'a483fa3d-7a97-48fa-8fe5-723336b546a9', 'name': 'volume-a483fa3d-7a97-48fa-8fe5-723336b546a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5d819bd2-f7f9-480d-83ae-fbd4d2e0d382', 'attached_at': '', 'detached_at': '', 'volume_id': 'a483fa3d-7a97-48fa-8fe5-723336b546a9', 'serial': 'a483fa3d-7a97-48fa-8fe5-723336b546a9'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 795.530580] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 795.532329] env[63372]: DEBUG nova.objects.base [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Object Instance<93a5d948-0629-4f53-a681-858d519acfa7> lazy-loaded attributes: flavor,info_cache {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 795.534506] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d6dbdc-990a-4d55-bfbf-512b1261bf9a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.541221] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "5819c38e-2cf9-4d16-b28a-5f23d35c3d44" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.541554] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "5819c38e-2cf9-4d16-b28a-5f23d35c3d44" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.541848] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "5819c38e-2cf9-4d16-b28a-5f23d35c3d44-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.542152] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "5819c38e-2cf9-4d16-b28a-5f23d35c3d44-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.542412] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "5819c38e-2cf9-4d16-b28a-5f23d35c3d44-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.548586] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 795.549291] env[63372]: INFO nova.compute.manager [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Terminating instance [ 795.551252] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-50f87627-e2e2-4ac5-920f-170092a974aa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.554296] env[63372]: DEBUG nova.compute.manager [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 795.554559] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 795.555689] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5b7b24-c79c-4a90-83aa-50b8f88605de {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.559956] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9583ee16-453e-4d16-822a-0431b21f8539 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "99f901a6-9bb3-4403-af0c-c8900f655cb3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.477s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.567257] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 795.571539] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-83382f96-802f-4708-afb1-c45dc1690263 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.580620] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024102, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.582480] env[63372]: DEBUG oslo_vmware.api [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 795.582480] env[63372]: value = "task-1024104" [ 795.582480] env[63372]: _type = "Task" [ 795.582480] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.593443] env[63372]: DEBUG oslo_vmware.api [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1024104, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.594412] env[63372]: DEBUG nova.network.neutron [req-8e5d9b1a-f840-4b04-871c-8a351e8db697 req-0c9e1588-5298-4e4f-b325-fcb5b0917063 service nova] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Updated VIF entry in instance network info cache for port a0cd67cc-8bd5-4d24-8620-042e1fecd124. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 795.594823] env[63372]: DEBUG nova.network.neutron [req-8e5d9b1a-f840-4b04-871c-8a351e8db697 req-0c9e1588-5298-4e4f-b325-fcb5b0917063 service nova] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Updating instance_info_cache with network_info: [{"id": "a0cd67cc-8bd5-4d24-8620-042e1fecd124", "address": "fa:16:3e:bb:a7:ca", "network": {"id": "8171761d-bd68-4efd-ad4f-e661badde0c4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1048168720-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1cc87c31eb77441daeec0fad5bb1ce87", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0cd67cc-8b", "ovs_interfaceid": "a0cd67cc-8bd5-4d24-8620-042e1fecd124", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.672382] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024098, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.596061} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.672695] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 7c9d551e-d210-4943-971f-b2829751fcfc/7c9d551e-d210-4943-971f-b2829751fcfc.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 795.672926] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 795.673201] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae638d9b-0a01-407e-8724-e9402d189644 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.680752] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 795.680752] env[63372]: value = "task-1024105" [ 795.680752] env[63372]: _type = "Task" [ 795.680752] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.689581] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024105, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.730725] env[63372]: DEBUG nova.compute.manager [req-b37baffc-b98b-4544-853b-929982f2b1e9 req-90f1f92b-7dae-4f12-9e25-f2f89b55edc9 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Received event network-vif-plugged-cd1ce4d9-46e7-41b8-8009-71283aa762b2 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.731835] env[63372]: DEBUG oslo_concurrency.lockutils [req-b37baffc-b98b-4544-853b-929982f2b1e9 req-90f1f92b-7dae-4f12-9e25-f2f89b55edc9 service nova] Acquiring lock "bbba9f28-045d-41ab-8539-5b2968fe3d54-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.731835] env[63372]: DEBUG oslo_concurrency.lockutils [req-b37baffc-b98b-4544-853b-929982f2b1e9 req-90f1f92b-7dae-4f12-9e25-f2f89b55edc9 service nova] Lock "bbba9f28-045d-41ab-8539-5b2968fe3d54-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.731835] env[63372]: DEBUG oslo_concurrency.lockutils [req-b37baffc-b98b-4544-853b-929982f2b1e9 req-90f1f92b-7dae-4f12-9e25-f2f89b55edc9 service nova] Lock "bbba9f28-045d-41ab-8539-5b2968fe3d54-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.731835] env[63372]: DEBUG nova.compute.manager [req-b37baffc-b98b-4544-853b-929982f2b1e9 req-90f1f92b-7dae-4f12-9e25-f2f89b55edc9 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] No waiting events found dispatching network-vif-plugged-cd1ce4d9-46e7-41b8-8009-71283aa762b2 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 795.731835] env[63372]: WARNING nova.compute.manager [req-b37baffc-b98b-4544-853b-929982f2b1e9 req-90f1f92b-7dae-4f12-9e25-f2f89b55edc9 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Received unexpected event network-vif-plugged-cd1ce4d9-46e7-41b8-8009-71283aa762b2 for instance with vm_state active and task_state None. [ 795.732940] env[63372]: DEBUG nova.compute.manager [req-b37baffc-b98b-4544-853b-929982f2b1e9 req-90f1f92b-7dae-4f12-9e25-f2f89b55edc9 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Received event network-changed-cd1ce4d9-46e7-41b8-8009-71283aa762b2 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.732940] env[63372]: DEBUG nova.compute.manager [req-b37baffc-b98b-4544-853b-929982f2b1e9 req-90f1f92b-7dae-4f12-9e25-f2f89b55edc9 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Refreshing instance network info cache due to event network-changed-cd1ce4d9-46e7-41b8-8009-71283aa762b2. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 795.732940] env[63372]: DEBUG oslo_concurrency.lockutils [req-b37baffc-b98b-4544-853b-929982f2b1e9 req-90f1f92b-7dae-4f12-9e25-f2f89b55edc9 service nova] Acquiring lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.732940] env[63372]: DEBUG oslo_concurrency.lockutils [req-b37baffc-b98b-4544-853b-929982f2b1e9 req-90f1f92b-7dae-4f12-9e25-f2f89b55edc9 service nova] Acquired lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.732940] env[63372]: DEBUG nova.network.neutron [req-b37baffc-b98b-4544-853b-929982f2b1e9 req-90f1f92b-7dae-4f12-9e25-f2f89b55edc9 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Refreshing network info cache for port cd1ce4d9-46e7-41b8-8009-71283aa762b2 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 795.753281] env[63372]: DEBUG nova.compute.utils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 795.754989] env[63372]: DEBUG nova.compute.manager [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 795.755192] env[63372]: DEBUG nova.network.neutron [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 795.767686] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024099, 'name': ReconfigVM_Task, 'duration_secs': 0.668459} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.768075] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Reconfigured VM instance instance-0000003f to attach disk [datastore1] de1b38a4-c7f2-420c-a050-7311976e4ca8/de1b38a4-c7f2-420c-a050-7311976e4ca8.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.769270] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3e1d4253-6d31-42cf-a2eb-2b4906c00f92 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.778729] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 795.778729] env[63372]: value = "task-1024106" [ 795.778729] env[63372]: _type = "Task" [ 795.778729] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.797316] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024106, 'name': Rename_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.832882] env[63372]: DEBUG nova.policy [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ddaa4503a23947a9a06962ed4d01aa32', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3d866ac9602470ea9f8d8159551acc8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 795.896257] env[63372]: INFO nova.compute.manager [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Rebuilding instance [ 795.940434] env[63372]: DEBUG nova.compute.manager [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 795.941383] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2b0a78-98d3-41d7-8052-e7e8997c3237 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.078426] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024102, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.080179] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 796.080494] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 796.080786] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Deleting the datastore file [datastore1] 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 796.081135] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4afe86b3-ee31-4702-9bee-f3b9a68e120f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.092741] env[63372]: DEBUG oslo_vmware.api [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1024104, 'name': PowerOffVM_Task, 'duration_secs': 0.40637} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.094107] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 796.094338] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 796.094779] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Waiting for the task: (returnval){ [ 796.094779] env[63372]: value = "task-1024107" [ 796.094779] env[63372]: _type = "Task" [ 796.094779] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.094923] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-407c1bd2-5165-437d-9ce9-6aa235ac251e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.099847] env[63372]: DEBUG oslo_concurrency.lockutils [req-8e5d9b1a-f840-4b04-871c-8a351e8db697 req-0c9e1588-5298-4e4f-b325-fcb5b0917063 service nova] Releasing lock "refresh_cache-7c9d551e-d210-4943-971f-b2829751fcfc" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.186508] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 796.186752] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 796.186962] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Deleting the datastore file [datastore2] 5819c38e-2cf9-4d16-b28a-5f23d35c3d44 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 796.188123] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d0e0c5bc-3c84-4a66-9334-1162e494b466 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.194551] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024105, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068879} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.195150] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.196027] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a38cb1-0ca0-4f81-9f79-2225dfb9918e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.200102] env[63372]: DEBUG oslo_vmware.api [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for the task: (returnval){ [ 796.200102] env[63372]: value = "task-1024109" [ 796.200102] env[63372]: _type = "Task" [ 796.200102] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.224918] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 7c9d551e-d210-4943-971f-b2829751fcfc/7c9d551e-d210-4943-971f-b2829751fcfc.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.228128] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b41a7ec-3140-452e-8e68-46daee56e1a4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.253853] env[63372]: DEBUG oslo_vmware.api [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1024109, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.261607] env[63372]: DEBUG nova.compute.manager [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 796.264140] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 796.264140] env[63372]: value = "task-1024110" [ 796.264140] env[63372]: _type = "Task" [ 796.264140] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.276369] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024110, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.291396] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024106, 'name': Rename_Task, 'duration_secs': 0.501221} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.291763] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 796.294178] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7620235-0461-44c9-a020-cde518ced944 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.302387] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 796.302387] env[63372]: value = "task-1024111" [ 796.302387] env[63372]: _type = "Task" [ 796.302387] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.314022] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024111, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.351252] env[63372]: DEBUG nova.network.neutron [-] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.452898] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 796.453224] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e98f2b8c-073d-4921-93d3-fc404b46cf9f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.464107] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 796.464107] env[63372]: value = "task-1024112" [ 796.464107] env[63372]: _type = "Task" [ 796.464107] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.475297] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024112, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.538961] env[63372]: DEBUG nova.network.neutron [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance_info_cache with network_info: [{"id": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "address": "fa:16:3e:3f:99:77", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c4c74ba-ad", "ovs_interfaceid": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.578962] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024102, 'name': CreateSnapshot_Task, 'duration_secs': 1.119795} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.579727] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 796.581366] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a2e83e-9df3-4cf4-968a-4770facc2de8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.623947] env[63372]: DEBUG oslo_vmware.api [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Task: {'id': task-1024107, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266056} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.624725] env[63372]: DEBUG nova.network.neutron [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Successfully created port: c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 796.627841] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 796.628070] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 796.628258] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 796.702261] env[63372]: DEBUG nova.network.neutron [req-b37baffc-b98b-4544-853b-929982f2b1e9 req-90f1f92b-7dae-4f12-9e25-f2f89b55edc9 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Updated VIF entry in instance network info cache for port cd1ce4d9-46e7-41b8-8009-71283aa762b2. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 796.702835] env[63372]: DEBUG nova.network.neutron [req-b37baffc-b98b-4544-853b-929982f2b1e9 req-90f1f92b-7dae-4f12-9e25-f2f89b55edc9 service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Updating instance_info_cache with network_info: [{"id": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "address": "fa:16:3e:17:c8:2e", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4cf27a3-fc", "ovs_interfaceid": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cd1ce4d9-46e7-41b8-8009-71283aa762b2", "address": "fa:16:3e:32:b0:d7", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd1ce4d9-46", "ovs_interfaceid": "cd1ce4d9-46e7-41b8-8009-71283aa762b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.711684] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Volume detach. Driver type: vmdk {{(pid=63372) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 796.711684] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41526b82-d649-4641-b23d-86b0bdcdd1d0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.722449] env[63372]: DEBUG oslo_vmware.api [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Task: {'id': task-1024109, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.309049} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.728256] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 796.728440] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 796.728683] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 796.729268] env[63372]: INFO nova.compute.manager [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Took 1.17 seconds to destroy the instance on the hypervisor. [ 796.729569] env[63372]: DEBUG oslo.service.loopingcall [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.731527] env[63372]: DEBUG nova.compute.manager [-] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 796.731635] env[63372]: DEBUG nova.network.neutron [-] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 796.737434] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e324670-67ab-41e7-af98-2fff50712839 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.790716] env[63372]: ERROR nova.compute.manager [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Failed to detach volume a483fa3d-7a97-48fa-8fe5-723336b546a9 from /dev/sda: nova.exception.InstanceNotFound: Instance 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382 could not be found. [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Traceback (most recent call last): [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] self.driver.rebuild(**kwargs) [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] raise NotImplementedError() [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] NotImplementedError [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] During handling of the above exception, another exception occurred: [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Traceback (most recent call last): [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] self.driver.detach_volume(context, old_connection_info, [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] return self._volumeops.detach_volume(connection_info, instance) [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] self._detach_volume_vmdk(connection_info, instance) [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] stable_ref.fetch_moref(session) [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] raise exception.InstanceNotFound(instance_id=self._uuid) [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] nova.exception.InstanceNotFound: Instance 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382 could not be found. [ 796.790716] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] [ 796.800450] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024110, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.814212] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024111, 'name': PowerOnVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.854878] env[63372]: INFO nova.compute.manager [-] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Took 1.34 seconds to deallocate network for instance. [ 796.881831] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8a6fb7-4fe3-48f5-9d8a-00e8da16b360 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.891210] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b6c57d-286a-4de6-bb05-6954cdc9b00e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.928989] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8892b3-6d95-4509-8379-a1dd601734a8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.939268] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f44fb2-e8b4-46f5-9bfd-0a5cd45ffbce {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.956775] env[63372]: DEBUG nova.compute.provider_tree [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 796.973707] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024112, 'name': PowerOffVM_Task, 'duration_secs': 0.274239} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.974022] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 796.974237] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 796.974991] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9147647-84d3-47e7-8629-c342f8fed0f1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.982952] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 796.983298] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e94c20af-728d-4638-9cb5-5784ca451314 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.010575] env[63372]: DEBUG nova.compute.utils [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Build of instance 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382 aborted: Failed to rebuild volume backed instance. {{(pid=63372) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 797.013397] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 797.013762] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 797.013762] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Deleting the datastore file [datastore1] f6675697-1529-46be-b28a-398ff3060d18 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.014767] env[63372]: ERROR nova.compute.manager [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382 aborted: Failed to rebuild volume backed instance. [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Traceback (most recent call last): [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] self.driver.rebuild(**kwargs) [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] raise NotImplementedError() [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] NotImplementedError [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] During handling of the above exception, another exception occurred: [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Traceback (most recent call last): [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] self._detach_root_volume(context, instance, root_bdm) [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] with excutils.save_and_reraise_exception(): [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] self.force_reraise() [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] raise self.value [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] self.driver.detach_volume(context, old_connection_info, [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] return self._volumeops.detach_volume(connection_info, instance) [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] self._detach_volume_vmdk(connection_info, instance) [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] stable_ref.fetch_moref(session) [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] raise exception.InstanceNotFound(instance_id=self._uuid) [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] nova.exception.InstanceNotFound: Instance 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382 could not be found. [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] During handling of the above exception, another exception occurred: [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Traceback (most recent call last): [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] yield [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 797.014767] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] self._do_rebuild_instance_with_claim( [ 797.016126] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 797.016126] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] self._do_rebuild_instance( [ 797.016126] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 797.016126] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] self._rebuild_default_impl(**kwargs) [ 797.016126] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 797.016126] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] self._rebuild_volume_backed_instance( [ 797.016126] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 797.016126] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] raise exception.BuildAbortException( [ 797.016126] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] nova.exception.BuildAbortException: Build of instance 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382 aborted: Failed to rebuild volume backed instance. [ 797.016126] env[63372]: ERROR nova.compute.manager [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] [ 797.017296] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7e9c3e3f-23d1-48b5-b374-0e609d29c534 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.027499] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 797.027499] env[63372]: value = "task-1024114" [ 797.027499] env[63372]: _type = "Task" [ 797.027499] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.037675] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024114, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.041505] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.123909] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 797.124074] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-efbf26f5-00cd-43d4-926f-c984997e3571 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.135901] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 797.135901] env[63372]: value = "task-1024115" [ 797.135901] env[63372]: _type = "Task" [ 797.135901] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.145971] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024115, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.211131] env[63372]: DEBUG oslo_concurrency.lockutils [req-b37baffc-b98b-4544-853b-929982f2b1e9 req-90f1f92b-7dae-4f12-9e25-f2f89b55edc9 service nova] Releasing lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.278841] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024110, 'name': ReconfigVM_Task, 'duration_secs': 0.829396} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.279299] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 7c9d551e-d210-4943-971f-b2829751fcfc/7c9d551e-d210-4943-971f-b2829751fcfc.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.280206] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8c8d845c-20d6-4f07-9a10-810bb4ff60b4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.289644] env[63372]: DEBUG nova.compute.manager [req-b2250b58-b223-4916-a553-9e23e4edc88e req-83214ced-b36a-42fd-97a9-59a8a4d311d4 service nova] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Received event network-vif-deleted-e018b1c1-a38b-4542-a2b4-bf87f12453ae {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 797.289950] env[63372]: INFO nova.compute.manager [req-b2250b58-b223-4916-a553-9e23e4edc88e req-83214ced-b36a-42fd-97a9-59a8a4d311d4 service nova] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Neutron deleted interface e018b1c1-a38b-4542-a2b4-bf87f12453ae; detaching it from the instance and deleting it from the info cache [ 797.290295] env[63372]: DEBUG nova.network.neutron [req-b2250b58-b223-4916-a553-9e23e4edc88e req-83214ced-b36a-42fd-97a9-59a8a4d311d4 service nova] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.293908] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 797.293908] env[63372]: value = "task-1024116" [ 797.293908] env[63372]: _type = "Task" [ 797.293908] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.301358] env[63372]: DEBUG nova.compute.manager [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 797.316368] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024116, 'name': Rename_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.327180] env[63372]: DEBUG oslo_vmware.api [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024111, 'name': PowerOnVM_Task, 'duration_secs': 0.833241} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.327180] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 797.327180] env[63372]: INFO nova.compute.manager [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Took 10.09 seconds to spawn the instance on the hypervisor. [ 797.327639] env[63372]: DEBUG nova.compute.manager [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 797.328766] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8ed754b-0535-463d-b0d4-908eb5d07d32 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.336455] env[63372]: DEBUG nova.virt.hardware [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 797.336851] env[63372]: DEBUG nova.virt.hardware [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 797.337124] env[63372]: DEBUG nova.virt.hardware [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 797.337440] env[63372]: DEBUG nova.virt.hardware [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 797.337712] env[63372]: DEBUG nova.virt.hardware [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 797.337986] env[63372]: DEBUG nova.virt.hardware [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 797.338356] env[63372]: DEBUG nova.virt.hardware [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 797.338644] env[63372]: DEBUG nova.virt.hardware [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 797.338950] env[63372]: DEBUG nova.virt.hardware [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 797.339344] env[63372]: DEBUG nova.virt.hardware [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 797.339456] env[63372]: DEBUG nova.virt.hardware [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 797.340620] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eccbc04-ad0e-4391-8691-27fc304f280b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.355310] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756a84be-ab37-45d5-839c-87abeb2a8523 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.365967] env[63372]: DEBUG oslo_concurrency.lockutils [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.481580] env[63372]: ERROR nova.scheduler.client.report [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [req-5c93ff66-4eec-446f-9b9b-b7173971ab2e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5c93ff66-4eec-446f-9b9b-b7173971ab2e"}]} [ 797.501401] env[63372]: DEBUG nova.scheduler.client.report [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Refreshing inventories for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 797.524073] env[63372]: DEBUG nova.scheduler.client.report [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Updating ProviderTree inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 184, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 797.524073] env[63372]: DEBUG nova.compute.provider_tree [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 184, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 797.546252] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024114, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139041} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.546252] env[63372]: DEBUG nova.scheduler.client.report [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Refreshing aggregate associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, aggregates: None {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 797.548174] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 797.548502] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.548732] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 797.548895] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 797.552008] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-388e550e-11b9-4017-b6e3-79d9308efb12 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.561235] env[63372]: DEBUG oslo_vmware.api [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 797.561235] env[63372]: value = "task-1024117" [ 797.561235] env[63372]: _type = "Task" [ 797.561235] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.572509] env[63372]: DEBUG nova.scheduler.client.report [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Refreshing trait associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 797.578957] env[63372]: DEBUG oslo_vmware.api [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024117, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.630596] env[63372]: DEBUG nova.network.neutron [-] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.648478] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024115, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.739912] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "interface-bbba9f28-045d-41ab-8539-5b2968fe3d54-cd1ce4d9-46e7-41b8-8009-71283aa762b2" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.740655] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-bbba9f28-045d-41ab-8539-5b2968fe3d54-cd1ce4d9-46e7-41b8-8009-71283aa762b2" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.795577] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3eeacdfd-b4bb-4e38-a089-cb2dcbbb8e86 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.808617] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024116, 'name': Rename_Task, 'duration_secs': 0.225444} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.812726] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 797.813354] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6a9b49a-3ee5-4283-aac3-08f0b5ea7bc6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.817761] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e2d328d-b5ad-4d65-b2d1-acd40001c5c9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.847448] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 797.847448] env[63372]: value = "task-1024118" [ 797.847448] env[63372]: _type = "Task" [ 797.847448] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.876711] env[63372]: DEBUG nova.compute.manager [req-ebc19dbe-c1c1-47d5-b915-de54a08a26c6 req-c406b2ba-c523-42e3-afb9-46701429be93 service nova] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Received event network-vif-deleted-6e151063-238b-4403-9365-be674f0638f5 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 797.877549] env[63372]: DEBUG nova.compute.manager [req-b2250b58-b223-4916-a553-9e23e4edc88e req-83214ced-b36a-42fd-97a9-59a8a4d311d4 service nova] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Detach interface failed, port_id=e018b1c1-a38b-4542-a2b4-bf87f12453ae, reason: Instance 5819c38e-2cf9-4d16-b28a-5f23d35c3d44 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 797.887870] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024118, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.888469] env[63372]: INFO nova.compute.manager [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Took 38.11 seconds to build instance. [ 798.072257] env[63372]: DEBUG oslo_vmware.api [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024117, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.103320] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329dd7bc-f2f6-4e6f-9134-5c0539f6c594 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.112169] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11fbddf3-51ad-4410-a2a7-af9aad394ff2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.142330] env[63372]: INFO nova.compute.manager [-] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Took 1.41 seconds to deallocate network for instance. [ 798.147607] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec213ea4-2d33-4491-9e41-4d685697faf6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.161713] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cece27-eabf-44ad-9438-c070512efda2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.165619] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024115, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.179829] env[63372]: DEBUG nova.compute.provider_tree [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 798.232709] env[63372]: DEBUG nova.network.neutron [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Successfully updated port: c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 798.247693] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.247693] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.248662] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d3c765-5325-4ac2-a087-b97f2b2d0da2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.272570] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a01ab23f-880b-43df-8ea2-b2e93fe763ff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.302718] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Reconfiguring VM to detach interface {{(pid=63372) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 798.303424] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a22e15b4-4bad-41b6-8d5a-0ef15bec9f32 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.323167] env[63372]: DEBUG oslo_vmware.api [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 798.323167] env[63372]: value = "task-1024119" [ 798.323167] env[63372]: _type = "Task" [ 798.323167] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.331849] env[63372]: DEBUG oslo_vmware.api [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024119, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.354897] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024118, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.394653] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f1a7207c-1156-49d7-893e-87e3cc152521 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "de1b38a4-c7f2-420c-a050-7311976e4ca8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.817s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.574971] env[63372]: DEBUG oslo_vmware.api [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024117, 'name': PowerOnVM_Task, 'duration_secs': 0.527722} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.575677] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 798.575677] env[63372]: DEBUG nova.compute.manager [None req-fbf74775-a5cc-4fd8-9dbd-1993a674e746 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 798.579321] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5f81c4-2645-48f4-9d18-cd62743c1edb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.593584] env[63372]: DEBUG nova.virt.hardware [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.593723] env[63372]: DEBUG nova.virt.hardware [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.593822] env[63372]: DEBUG nova.virt.hardware [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.594013] env[63372]: DEBUG nova.virt.hardware [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.594170] env[63372]: DEBUG nova.virt.hardware [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.594322] env[63372]: DEBUG nova.virt.hardware [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.594544] env[63372]: DEBUG nova.virt.hardware [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.594758] env[63372]: DEBUG nova.virt.hardware [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.594954] env[63372]: DEBUG nova.virt.hardware [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.595146] env[63372]: DEBUG nova.virt.hardware [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.595384] env[63372]: DEBUG nova.virt.hardware [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.598841] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74fda47-c67f-4ef4-a11c-94a80b95f374 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.603064] env[63372]: DEBUG oslo_concurrency.lockutils [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "f0c60559-c072-4b61-afe8-03d6c131b307" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.604020] env[63372]: DEBUG oslo_concurrency.lockutils [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "f0c60559-c072-4b61-afe8-03d6c131b307" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.604020] env[63372]: DEBUG oslo_concurrency.lockutils [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "f0c60559-c072-4b61-afe8-03d6c131b307-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.604020] env[63372]: DEBUG oslo_concurrency.lockutils [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "f0c60559-c072-4b61-afe8-03d6c131b307-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.604020] env[63372]: DEBUG oslo_concurrency.lockutils [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "f0c60559-c072-4b61-afe8-03d6c131b307-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.606267] env[63372]: INFO nova.compute.manager [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Terminating instance [ 798.608340] env[63372]: DEBUG nova.compute.manager [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 798.608906] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 798.609577] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c1579e-4286-45b3-8c8a-f74f596cf028 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.616292] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc57cd63-f5d1-43f6-bb9f-ed68c00ed08e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.624900] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 798.625542] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b2997c6-747e-4a56-94fc-e1fd1c9b2097 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.635587] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Instance VIF info [] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 798.641988] env[63372]: DEBUG oslo.service.loopingcall [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.644477] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6675697-1529-46be-b28a-398ff3060d18] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 798.645093] env[63372]: DEBUG oslo_vmware.api [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 798.645093] env[63372]: value = "task-1024120" [ 798.645093] env[63372]: _type = "Task" [ 798.645093] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.645329] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da157bfa-e358-4e44-b634-990d9d63df84 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.661682] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.671237] env[63372]: DEBUG oslo_vmware.api [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024120, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.675170] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 798.675170] env[63372]: value = "task-1024121" [ 798.675170] env[63372]: _type = "Task" [ 798.675170] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.675746] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024115, 'name': CloneVM_Task} progress is 95%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.687038] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024121, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.717252] env[63372]: DEBUG nova.scheduler.client.report [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 88 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 798.717552] env[63372]: DEBUG nova.compute.provider_tree [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 88 to 89 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 798.717739] env[63372]: DEBUG nova.compute.provider_tree [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 798.737817] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.738165] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.738165] env[63372]: DEBUG nova.network.neutron [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 798.834207] env[63372]: DEBUG oslo_vmware.api [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024119, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.855264] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024118, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.035064] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.156274] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024115, 'name': CloneVM_Task, 'duration_secs': 1.885076} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.156561] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Created linked-clone VM from snapshot [ 799.157386] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867e9e00-1602-4131-89aa-567d49a9df39 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.167778] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Uploading image de1edf71-70ee-4a4d-9486-cbfcce731c28 {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 799.175454] env[63372]: DEBUG oslo_vmware.api [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024120, 'name': PowerOffVM_Task, 'duration_secs': 0.298207} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.175819] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 799.175941] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 799.176281] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1c1aaad-1fdb-4fe8-b9b1-982fa8401c10 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.187169] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024121, 'name': CreateVM_Task, 'duration_secs': 0.430194} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.189345] env[63372]: DEBUG oslo_vmware.rw_handles [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 799.189345] env[63372]: value = "vm-227360" [ 799.189345] env[63372]: _type = "VirtualMachine" [ 799.189345] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 799.189544] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6675697-1529-46be-b28a-398ff3060d18] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 799.190086] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-850b5820-ede1-4926-8945-467d884fcea9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.191781] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.191970] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.192312] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 799.192557] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2676f68e-dd8d-4d45-8036-48f211d0c849 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.197453] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 799.197453] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52c6697f-d775-4ad2-f42a-4a7f588abb73" [ 799.197453] env[63372]: _type = "Task" [ 799.197453] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.202440] env[63372]: DEBUG oslo_vmware.rw_handles [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lease: (returnval){ [ 799.202440] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5209ae7b-e71b-7d23-436b-37c58572567b" [ 799.202440] env[63372]: _type = "HttpNfcLease" [ 799.202440] env[63372]: } obtained for exporting VM: (result){ [ 799.202440] env[63372]: value = "vm-227360" [ 799.202440] env[63372]: _type = "VirtualMachine" [ 799.202440] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 799.202880] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the lease: (returnval){ [ 799.202880] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5209ae7b-e71b-7d23-436b-37c58572567b" [ 799.202880] env[63372]: _type = "HttpNfcLease" [ 799.202880] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 799.210374] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c6697f-d775-4ad2-f42a-4a7f588abb73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.217047] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 799.217047] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5209ae7b-e71b-7d23-436b-37c58572567b" [ 799.217047] env[63372]: _type = "HttpNfcLease" [ 799.217047] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 799.217380] env[63372]: DEBUG oslo_vmware.rw_handles [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 799.217380] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5209ae7b-e71b-7d23-436b-37c58572567b" [ 799.217380] env[63372]: _type = "HttpNfcLease" [ 799.217380] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 799.218170] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35be4e7c-a8e1-4087-b80a-97bc2d1b864e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.229524] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.985s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.230166] env[63372]: DEBUG nova.compute.manager [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 799.232963] env[63372]: DEBUG oslo_vmware.rw_handles [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ddbbba-ec2c-fc0d-54ae-3f9bb9b2fa1e/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 799.233191] env[63372]: DEBUG oslo_vmware.rw_handles [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ddbbba-ec2c-fc0d-54ae-3f9bb9b2fa1e/disk-0.vmdk for reading. {{(pid=63372) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 799.234893] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.437s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.236499] env[63372]: INFO nova.compute.claims [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.326291] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 799.326519] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 799.328661] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Deleting the datastore file [datastore1] f0c60559-c072-4b61-afe8-03d6c131b307 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 799.333935] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-edfa3fb8-3356-4791-a42c-79535709b405 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.343617] env[63372]: DEBUG oslo_vmware.api [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024119, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.344535] env[63372]: DEBUG oslo_vmware.api [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 799.344535] env[63372]: value = "task-1024124" [ 799.344535] env[63372]: _type = "Task" [ 799.344535] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.358712] env[63372]: DEBUG oslo_vmware.api [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.362427] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024118, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.368588] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b9edca9a-3d28-40b0-8b22-4112fe2263b2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.370135] env[63372]: DEBUG nova.network.neutron [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 799.541768] env[63372]: DEBUG nova.network.neutron [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updating instance_info_cache with network_info: [{"id": "c92e8cc9-dc72-4f20-b087-1d323c502108", "address": "fa:16:3e:39:9e:be", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc92e8cc9-dc", "ovs_interfaceid": "c92e8cc9-dc72-4f20-b087-1d323c502108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.709476] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c6697f-d775-4ad2-f42a-4a7f588abb73, 'name': SearchDatastore_Task, 'duration_secs': 0.013048} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.709952] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.710226] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.710532] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.710802] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.711635] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.712107] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3659d56e-9a7c-4fb8-9e30-25ee5fed2d5c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.726521] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.726728] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 799.728041] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20f7e436-acc6-4354-a32c-42cc7c902e47 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.736225] env[63372]: DEBUG nova.compute.utils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 799.738410] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 799.738410] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521b526c-7926-ca50-9b00-cfb1e7c7bb51" [ 799.738410] env[63372]: _type = "Task" [ 799.738410] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.739721] env[63372]: DEBUG nova.compute.manager [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 799.740408] env[63372]: DEBUG nova.network.neutron [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 799.758995] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521b526c-7926-ca50-9b00-cfb1e7c7bb51, 'name': SearchDatastore_Task, 'duration_secs': 0.013049} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.760489] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb9256a0-adf4-45cb-819f-f79ddc14d2d9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.768519] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 799.768519] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52517dd9-ee0f-60ab-a4ea-bb68bf158b15" [ 799.768519] env[63372]: _type = "Task" [ 799.768519] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.780556] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52517dd9-ee0f-60ab-a4ea-bb68bf158b15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.793245] env[63372]: DEBUG nova.policy [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '053f2ae589f5418785bbb783b635e0a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a14a1aa8a0574b65a312754ae1091f37', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 799.840183] env[63372]: DEBUG oslo_vmware.api [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024119, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.871740] env[63372]: DEBUG oslo_vmware.api [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024124, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.260574} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.872199] env[63372]: DEBUG oslo_vmware.api [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024118, 'name': PowerOnVM_Task, 'duration_secs': 1.659693} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.873990] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 799.874497] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 799.874497] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 799.874639] env[63372]: INFO nova.compute.manager [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Took 1.27 seconds to destroy the instance on the hypervisor. [ 799.874965] env[63372]: DEBUG oslo.service.loopingcall [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.877273] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 799.877273] env[63372]: INFO nova.compute.manager [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Took 10.00 seconds to spawn the instance on the hypervisor. [ 799.877273] env[63372]: DEBUG nova.compute.manager [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.877508] env[63372]: DEBUG nova.compute.manager [-] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 799.877644] env[63372]: DEBUG nova.network.neutron [-] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 799.880998] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c1e043-c9ca-4cba-9531-b50c81f90891 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.045599] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.046161] env[63372]: DEBUG nova.compute.manager [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Instance network_info: |[{"id": "c92e8cc9-dc72-4f20-b087-1d323c502108", "address": "fa:16:3e:39:9e:be", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc92e8cc9-dc", "ovs_interfaceid": "c92e8cc9-dc72-4f20-b087-1d323c502108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 800.049214] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:9e:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c92e8cc9-dc72-4f20-b087-1d323c502108', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.067459] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Creating folder: Project (f3d866ac9602470ea9f8d8159551acc8). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 800.067459] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-43571e2f-16d2-41a6-93b9-50a2db1dbc53 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.089233] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Created folder: Project (f3d866ac9602470ea9f8d8159551acc8) in parent group-v227230. [ 800.089233] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Creating folder: Instances. Parent ref: group-v227362. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 800.089233] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0add46b1-2e37-47bf-ae80-594b9eabf53a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.100395] env[63372]: DEBUG nova.network.neutron [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Successfully created port: bf61306f-1264-4480-b485-299826369f7d {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 800.103824] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Created folder: Instances in parent group-v227362. [ 800.104316] env[63372]: DEBUG oslo.service.loopingcall [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.104658] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 800.105132] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1aaec3c5-70d7-41fc-8e5f-b791054328ef {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.128310] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.128310] env[63372]: value = "task-1024127" [ 800.128310] env[63372]: _type = "Task" [ 800.128310] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.138644] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024127, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.232034] env[63372]: DEBUG oslo_concurrency.lockutils [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Acquiring lock "5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.232034] env[63372]: DEBUG oslo_concurrency.lockutils [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Lock "5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.232034] env[63372]: DEBUG oslo_concurrency.lockutils [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Acquiring lock "5d819bd2-f7f9-480d-83ae-fbd4d2e0d382-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.232034] env[63372]: DEBUG oslo_concurrency.lockutils [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Lock "5d819bd2-f7f9-480d-83ae-fbd4d2e0d382-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.232034] env[63372]: DEBUG oslo_concurrency.lockutils [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Lock "5d819bd2-f7f9-480d-83ae-fbd4d2e0d382-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.232813] env[63372]: INFO nova.compute.manager [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Terminating instance [ 800.235665] env[63372]: DEBUG nova.compute.manager [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 800.236187] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7e8edbbc-8427-4add-a78a-d51568b7a5f1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.239281] env[63372]: DEBUG nova.compute.manager [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 800.258203] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3ee9fb-08d1-4105-9bac-db41baa1a5ab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.288888] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52517dd9-ee0f-60ab-a4ea-bb68bf158b15, 'name': SearchDatastore_Task, 'duration_secs': 0.016014} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.288888] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.288888] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] f6675697-1529-46be-b28a-398ff3060d18/f6675697-1529-46be-b28a-398ff3060d18.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 800.288888] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d244da0a-4ead-4f22-933a-ebffd3e70a9f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.311680] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 800.311680] env[63372]: value = "task-1024128" [ 800.311680] env[63372]: _type = "Task" [ 800.311680] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.312117] env[63372]: WARNING nova.virt.vmwareapi.driver [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382 could not be found. [ 800.312408] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 800.317019] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eae6a2c3-0576-4ebf-9909-31db78507591 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.326595] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024128, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.342724] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d534ecf5-4f96-497e-b466-2ada288538b1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.368374] env[63372]: DEBUG oslo_vmware.api [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024119, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.398419] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382 could not be found. [ 800.398683] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 800.398902] env[63372]: INFO nova.compute.manager [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Took 0.16 seconds to destroy the instance on the hypervisor. [ 800.399312] env[63372]: DEBUG oslo.service.loopingcall [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.402169] env[63372]: DEBUG nova.compute.manager [-] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 800.402291] env[63372]: DEBUG nova.network.neutron [-] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 800.407362] env[63372]: INFO nova.compute.manager [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Took 37.85 seconds to build instance. [ 800.423774] env[63372]: DEBUG nova.compute.manager [req-bf2464a0-c652-4c64-937d-3e8fc55f3be2 req-e38a1399-f3d6-4ee8-920a-3a6cacb80691 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Received event network-vif-plugged-c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.424864] env[63372]: DEBUG oslo_concurrency.lockutils [req-bf2464a0-c652-4c64-937d-3e8fc55f3be2 req-e38a1399-f3d6-4ee8-920a-3a6cacb80691 service nova] Acquiring lock "c01a5d24-eb46-4a69-993e-753880ce8e85-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.424864] env[63372]: DEBUG oslo_concurrency.lockutils [req-bf2464a0-c652-4c64-937d-3e8fc55f3be2 req-e38a1399-f3d6-4ee8-920a-3a6cacb80691 service nova] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.424864] env[63372]: DEBUG oslo_concurrency.lockutils [req-bf2464a0-c652-4c64-937d-3e8fc55f3be2 req-e38a1399-f3d6-4ee8-920a-3a6cacb80691 service nova] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.425014] env[63372]: DEBUG nova.compute.manager [req-bf2464a0-c652-4c64-937d-3e8fc55f3be2 req-e38a1399-f3d6-4ee8-920a-3a6cacb80691 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] No waiting events found dispatching network-vif-plugged-c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 800.425178] env[63372]: WARNING nova.compute.manager [req-bf2464a0-c652-4c64-937d-3e8fc55f3be2 req-e38a1399-f3d6-4ee8-920a-3a6cacb80691 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Received unexpected event network-vif-plugged-c92e8cc9-dc72-4f20-b087-1d323c502108 for instance with vm_state building and task_state spawning. [ 800.425384] env[63372]: DEBUG nova.compute.manager [req-bf2464a0-c652-4c64-937d-3e8fc55f3be2 req-e38a1399-f3d6-4ee8-920a-3a6cacb80691 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Received event network-changed-c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.425609] env[63372]: DEBUG nova.compute.manager [req-bf2464a0-c652-4c64-937d-3e8fc55f3be2 req-e38a1399-f3d6-4ee8-920a-3a6cacb80691 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Refreshing instance network info cache due to event network-changed-c92e8cc9-dc72-4f20-b087-1d323c502108. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 800.425971] env[63372]: DEBUG oslo_concurrency.lockutils [req-bf2464a0-c652-4c64-937d-3e8fc55f3be2 req-e38a1399-f3d6-4ee8-920a-3a6cacb80691 service nova] Acquiring lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.426140] env[63372]: DEBUG oslo_concurrency.lockutils [req-bf2464a0-c652-4c64-937d-3e8fc55f3be2 req-e38a1399-f3d6-4ee8-920a-3a6cacb80691 service nova] Acquired lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.426377] env[63372]: DEBUG nova.network.neutron [req-bf2464a0-c652-4c64-937d-3e8fc55f3be2 req-e38a1399-f3d6-4ee8-920a-3a6cacb80691 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Refreshing network info cache for port c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 800.640128] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024127, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.832154] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024128, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.852313] env[63372]: DEBUG oslo_vmware.api [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024119, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.865374] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b851da9f-bf2d-4fa8-af37-33d700e722e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.877977] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a80b3d4-7790-456c-a593-f40ff5feda9d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.915920] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c82a598c-f37f-4de4-a4cd-34cd7068e2e1 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "7c9d551e-d210-4943-971f-b2829751fcfc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.770s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.917526] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92efb3a7-c437-4678-8a3f-3f27b356395c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.927991] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab286e73-f008-4484-9a59-f4fad1978acd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.946995] env[63372]: DEBUG nova.compute.provider_tree [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.950460] env[63372]: DEBUG nova.compute.manager [req-842af859-107c-4234-9c9f-26d5287aaec5 req-32817a8a-d767-4999-8579-6200886f8798 service nova] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Received event network-vif-deleted-c4e27cbd-df52-4e92-9621-3b068191a460 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.950745] env[63372]: INFO nova.compute.manager [req-842af859-107c-4234-9c9f-26d5287aaec5 req-32817a8a-d767-4999-8579-6200886f8798 service nova] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Neutron deleted interface c4e27cbd-df52-4e92-9621-3b068191a460; detaching it from the instance and deleting it from the info cache [ 800.951365] env[63372]: DEBUG nova.network.neutron [req-842af859-107c-4234-9c9f-26d5287aaec5 req-32817a8a-d767-4999-8579-6200886f8798 service nova] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.063198] env[63372]: DEBUG nova.network.neutron [-] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.140515] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024127, 'name': CreateVM_Task, 'duration_secs': 0.840567} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.140948] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 801.141975] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.142408] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.142990] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 801.143382] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bb27d05-faf3-4b13-801d-646a5fffcc38 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.152030] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 801.152030] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523a6dfd-8c61-b7f2-ecbe-1f56285956f8" [ 801.152030] env[63372]: _type = "Task" [ 801.152030] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.159658] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523a6dfd-8c61-b7f2-ecbe-1f56285956f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.254447] env[63372]: DEBUG nova.compute.manager [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 801.282251] env[63372]: DEBUG nova.virt.hardware [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 801.282889] env[63372]: DEBUG nova.virt.hardware [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 801.283081] env[63372]: DEBUG nova.virt.hardware [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.283832] env[63372]: DEBUG nova.virt.hardware [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 801.284100] env[63372]: DEBUG nova.virt.hardware [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.284361] env[63372]: DEBUG nova.virt.hardware [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 801.284735] env[63372]: DEBUG nova.virt.hardware [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 801.284950] env[63372]: DEBUG nova.virt.hardware [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 801.285197] env[63372]: DEBUG nova.virt.hardware [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 801.285583] env[63372]: DEBUG nova.virt.hardware [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 801.286091] env[63372]: DEBUG nova.virt.hardware [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.286772] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc94aa94-8a76-4b8e-bfed-a68ec33fdca5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.297313] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b12602c-eb00-4cb7-a3cc-939f19210c97 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.325377] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024128, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.780408} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.325719] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] f6675697-1529-46be-b28a-398ff3060d18/f6675697-1529-46be-b28a-398ff3060d18.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 801.326089] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.326446] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-88e1c361-9338-41b9-b77a-5edcf8112a76 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.338918] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 801.338918] env[63372]: value = "task-1024129" [ 801.338918] env[63372]: _type = "Task" [ 801.338918] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.349371] env[63372]: DEBUG oslo_vmware.api [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024119, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.354961] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024129, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.416225] env[63372]: DEBUG nova.network.neutron [req-bf2464a0-c652-4c64-937d-3e8fc55f3be2 req-e38a1399-f3d6-4ee8-920a-3a6cacb80691 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updated VIF entry in instance network info cache for port c92e8cc9-dc72-4f20-b087-1d323c502108. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 801.416731] env[63372]: DEBUG nova.network.neutron [req-bf2464a0-c652-4c64-937d-3e8fc55f3be2 req-e38a1399-f3d6-4ee8-920a-3a6cacb80691 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updating instance_info_cache with network_info: [{"id": "c92e8cc9-dc72-4f20-b087-1d323c502108", "address": "fa:16:3e:39:9e:be", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc92e8cc9-dc", "ovs_interfaceid": "c92e8cc9-dc72-4f20-b087-1d323c502108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.454446] env[63372]: DEBUG nova.scheduler.client.report [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 801.461023] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d905c021-3237-473b-abc6-bc43d10200dd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.473883] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19578c74-acef-4050-8408-bcd92e29852d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.486967] env[63372]: DEBUG nova.network.neutron [-] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.519383] env[63372]: DEBUG nova.compute.manager [req-842af859-107c-4234-9c9f-26d5287aaec5 req-32817a8a-d767-4999-8579-6200886f8798 service nova] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Detach interface failed, port_id=c4e27cbd-df52-4e92-9621-3b068191a460, reason: Instance f0c60559-c072-4b61-afe8-03d6c131b307 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 801.566686] env[63372]: INFO nova.compute.manager [-] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Took 1.69 seconds to deallocate network for instance. [ 801.664325] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523a6dfd-8c61-b7f2-ecbe-1f56285956f8, 'name': SearchDatastore_Task, 'duration_secs': 0.023971} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.664664] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.664940] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 801.665200] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.665349] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.665554] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 801.665837] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c12b02fe-2d18-4098-b94d-fb2a384ac9e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.688040] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.688040] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 801.688821] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2aaff128-22e6-4a13-90c8-4ca1bd29ae87 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.695049] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 801.695049] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5254cdbf-f81b-f358-d201-664db424fe66" [ 801.695049] env[63372]: _type = "Task" [ 801.695049] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.703897] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5254cdbf-f81b-f358-d201-664db424fe66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.849917] env[63372]: DEBUG oslo_vmware.api [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024119, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.854289] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024129, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.166643} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.854289] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 801.854872] env[63372]: DEBUG nova.network.neutron [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Successfully updated port: bf61306f-1264-4480-b485-299826369f7d {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 801.856586] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d8cb2e-aac3-42a9-85a8-8afa7a58df48 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.887360] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] f6675697-1529-46be-b28a-398ff3060d18/f6675697-1529-46be-b28a-398ff3060d18.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.888467] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5a3db3e-a592-4722-bcf0-035cebd6e11f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.911619] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 801.911619] env[63372]: value = "task-1024130" [ 801.911619] env[63372]: _type = "Task" [ 801.911619] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.919325] env[63372]: DEBUG oslo_concurrency.lockutils [req-bf2464a0-c652-4c64-937d-3e8fc55f3be2 req-e38a1399-f3d6-4ee8-920a-3a6cacb80691 service nova] Releasing lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.923847] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024130, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.959659] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.725s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.960310] env[63372]: DEBUG nova.compute.manager [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 801.967022] env[63372]: DEBUG oslo_concurrency.lockutils [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.618s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.967022] env[63372]: DEBUG nova.objects.instance [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Lazy-loading 'resources' on Instance uuid 4c9eb955-3bed-4b26-866f-b2a876b835dd {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 801.988299] env[63372]: INFO nova.compute.manager [-] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Took 1.59 seconds to deallocate network for instance. [ 802.076918] env[63372]: DEBUG oslo_concurrency.lockutils [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.207259] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5254cdbf-f81b-f358-d201-664db424fe66, 'name': SearchDatastore_Task, 'duration_secs': 0.021496} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.208320] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bb638ba-a30f-46cc-b04c-1eedc9ef41fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.215708] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 802.215708] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52bd2e57-ee45-dd82-c14c-0577e818761d" [ 802.215708] env[63372]: _type = "Task" [ 802.215708] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.225875] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52bd2e57-ee45-dd82-c14c-0577e818761d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.346235] env[63372]: DEBUG oslo_vmware.api [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024119, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.360142] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "refresh_cache-357505d0-f306-4e11-8a62-e03cfab2b7c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.360312] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquired lock "refresh_cache-357505d0-f306-4e11-8a62-e03cfab2b7c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.360472] env[63372]: DEBUG nova.network.neutron [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 802.425209] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024130, 'name': ReconfigVM_Task, 'duration_secs': 0.423971} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.425563] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Reconfigured VM instance instance-0000003d to attach disk [datastore2] f6675697-1529-46be-b28a-398ff3060d18/f6675697-1529-46be-b28a-398ff3060d18.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.427226] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-74499b5f-a0cc-4b6f-9a3b-73be18277646 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.436660] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 802.436660] env[63372]: value = "task-1024131" [ 802.436660] env[63372]: _type = "Task" [ 802.436660] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.452168] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024131, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.466843] env[63372]: DEBUG nova.compute.utils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 802.472846] env[63372]: DEBUG nova.compute.manager [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 802.473113] env[63372]: DEBUG nova.network.neutron [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 802.562564] env[63372]: DEBUG nova.policy [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f8e364e3aed40248cc0d11741f76154', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b4548b52cd704cc0b054e2f7d1562b9a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 802.577213] env[63372]: INFO nova.compute.manager [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Took 0.59 seconds to detach 1 volumes for instance. [ 802.579670] env[63372]: DEBUG nova.compute.manager [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Deleting volume: a483fa3d-7a97-48fa-8fe5-723336b546a9 {{(pid=63372) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 802.727957] env[63372]: DEBUG nova.compute.manager [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Received event network-vif-deleted-62ba51bf-2e6c-4205-93c0-e45bfd98f6ae {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.727957] env[63372]: DEBUG nova.compute.manager [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Received event network-vif-plugged-bf61306f-1264-4480-b485-299826369f7d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.727957] env[63372]: DEBUG oslo_concurrency.lockutils [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] Acquiring lock "357505d0-f306-4e11-8a62-e03cfab2b7c5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.727957] env[63372]: DEBUG oslo_concurrency.lockutils [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] Lock "357505d0-f306-4e11-8a62-e03cfab2b7c5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.727957] env[63372]: DEBUG oslo_concurrency.lockutils [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] Lock "357505d0-f306-4e11-8a62-e03cfab2b7c5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.727957] env[63372]: DEBUG nova.compute.manager [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] No waiting events found dispatching network-vif-plugged-bf61306f-1264-4480-b485-299826369f7d {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 802.727957] env[63372]: WARNING nova.compute.manager [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Received unexpected event network-vif-plugged-bf61306f-1264-4480-b485-299826369f7d for instance with vm_state building and task_state spawning. [ 802.727957] env[63372]: DEBUG nova.compute.manager [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Received event network-changed-bf61306f-1264-4480-b485-299826369f7d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.727957] env[63372]: DEBUG nova.compute.manager [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Refreshing instance network info cache due to event network-changed-bf61306f-1264-4480-b485-299826369f7d. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 802.727957] env[63372]: DEBUG oslo_concurrency.lockutils [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] Acquiring lock "refresh_cache-357505d0-f306-4e11-8a62-e03cfab2b7c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.739914] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52bd2e57-ee45-dd82-c14c-0577e818761d, 'name': SearchDatastore_Task, 'duration_secs': 0.01451} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.740105] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.740374] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] c01a5d24-eb46-4a69-993e-753880ce8e85/c01a5d24-eb46-4a69-993e-753880ce8e85.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 802.740653] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-698f5eba-2116-4cc8-bcaf-ef583bcdff57 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.752173] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 802.752173] env[63372]: value = "task-1024133" [ 802.752173] env[63372]: _type = "Task" [ 802.752173] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.762867] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024133, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.848024] env[63372]: DEBUG oslo_vmware.api [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024119, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.940522] env[63372]: DEBUG nova.network.neutron [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 802.953820] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024131, 'name': Rename_Task, 'duration_secs': 0.20771} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.954118] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 802.959671] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a3c2c345-e3fd-4cf7-948c-75ede7d41727 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.969866] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 802.969866] env[63372]: value = "task-1024134" [ 802.969866] env[63372]: _type = "Task" [ 802.969866] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.976369] env[63372]: DEBUG nova.compute.manager [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 802.986691] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024134, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.046770] env[63372]: DEBUG nova.network.neutron [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Successfully created port: c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.068154] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d62d2a1-d841-4d1e-9f22-cae4d847a031 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.077379] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce1be9fa-b97a-4295-a0ab-6897e234971a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.111530] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52838c70-bf60-425e-a21a-98a7fde045cf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.124287] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ae158c-1661-44f9-8871-c2b46906f4bd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.134304] env[63372]: DEBUG oslo_concurrency.lockutils [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.144554] env[63372]: DEBUG nova.compute.provider_tree [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.237431] env[63372]: DEBUG nova.network.neutron [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Updating instance_info_cache with network_info: [{"id": "bf61306f-1264-4480-b485-299826369f7d", "address": "fa:16:3e:13:43:f9", "network": {"id": "2f241a45-cc18-46f6-b0a3-1aad8d2713fd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1953733697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a14a1aa8a0574b65a312754ae1091f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a88b707-352e-4be7-b1d6-ad6074b40ed9", "external-id": "nsx-vlan-transportzone-789", "segmentation_id": 789, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf61306f-12", "ovs_interfaceid": "bf61306f-1264-4480-b485-299826369f7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.267058] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024133, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.352657] env[63372]: DEBUG oslo_vmware.api [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024119, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.490033] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024134, 'name': PowerOnVM_Task} progress is 87%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.647860] env[63372]: DEBUG nova.scheduler.client.report [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 803.740826] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Releasing lock "refresh_cache-357505d0-f306-4e11-8a62-e03cfab2b7c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.741027] env[63372]: DEBUG nova.compute.manager [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Instance network_info: |[{"id": "bf61306f-1264-4480-b485-299826369f7d", "address": "fa:16:3e:13:43:f9", "network": {"id": "2f241a45-cc18-46f6-b0a3-1aad8d2713fd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1953733697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a14a1aa8a0574b65a312754ae1091f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a88b707-352e-4be7-b1d6-ad6074b40ed9", "external-id": "nsx-vlan-transportzone-789", "segmentation_id": 789, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf61306f-12", "ovs_interfaceid": "bf61306f-1264-4480-b485-299826369f7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 803.742176] env[63372]: DEBUG oslo_concurrency.lockutils [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] Acquired lock "refresh_cache-357505d0-f306-4e11-8a62-e03cfab2b7c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.742176] env[63372]: DEBUG nova.network.neutron [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Refreshing network info cache for port bf61306f-1264-4480-b485-299826369f7d {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 803.743684] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:43:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0a88b707-352e-4be7-b1d6-ad6074b40ed9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bf61306f-1264-4480-b485-299826369f7d', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 803.753168] env[63372]: DEBUG oslo.service.loopingcall [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 803.754261] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 803.754929] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2b95d850-ec59-44d7-a643-dac3a8228e53 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.780014] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024133, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.71839} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.782059] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] c01a5d24-eb46-4a69-993e-753880ce8e85/c01a5d24-eb46-4a69-993e-753880ce8e85.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 803.782059] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.782189] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 803.782189] env[63372]: value = "task-1024135" [ 803.782189] env[63372]: _type = "Task" [ 803.782189] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.782429] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5e6670e-51a4-4d00-909e-831ccbbb1bb9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.793680] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024135, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.795095] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 803.795095] env[63372]: value = "task-1024136" [ 803.795095] env[63372]: _type = "Task" [ 803.795095] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.807777] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024136, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.851544] env[63372]: DEBUG oslo_vmware.api [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024119, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.982854] env[63372]: DEBUG oslo_vmware.api [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024134, 'name': PowerOnVM_Task, 'duration_secs': 0.797614} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.983180] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 803.983389] env[63372]: DEBUG nova.compute.manager [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 803.984209] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f2d278-8773-4e66-8752-8cec0740c4e9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.993161] env[63372]: DEBUG nova.compute.manager [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 804.038170] env[63372]: DEBUG nova.virt.hardware [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='a62a84c4040dc2aadeda3b74a4826da2',container_format='bare',created_at=2024-09-30T11:38:30Z,direct_url=,disk_format='vmdk',id=3c9be794-7dcb-43df-a9b2-d9289ac8889f,min_disk=1,min_ram=0,name='tempest-test-snap-13044637',owner='b4548b52cd704cc0b054e2f7d1562b9a',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-09-30T11:38:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.038324] env[63372]: DEBUG nova.virt.hardware [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.038430] env[63372]: DEBUG nova.virt.hardware [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.038631] env[63372]: DEBUG nova.virt.hardware [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.038777] env[63372]: DEBUG nova.virt.hardware [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.039058] env[63372]: DEBUG nova.virt.hardware [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.039160] env[63372]: DEBUG nova.virt.hardware [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.039327] env[63372]: DEBUG nova.virt.hardware [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.039546] env[63372]: DEBUG nova.virt.hardware [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.039720] env[63372]: DEBUG nova.virt.hardware [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.040461] env[63372]: DEBUG nova.virt.hardware [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.041709] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd19d3b8-ddb8-4ff5-b7df-00835622246d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.052496] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4729184e-cde5-4e38-aadf-faabf6c75e5b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.155352] env[63372]: DEBUG oslo_concurrency.lockutils [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.192s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.158614] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.603s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.159849] env[63372]: DEBUG nova.objects.instance [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lazy-loading 'resources' on Instance uuid febeca1c-3370-4c74-8bd9-efad313df1e2 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 804.191990] env[63372]: INFO nova.scheduler.client.report [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Deleted allocations for instance 4c9eb955-3bed-4b26-866f-b2a876b835dd [ 804.297262] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024135, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.309526] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024136, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079448} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.312433] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 804.313263] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8718660e-61d2-494d-bb77-5d7735506b5b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.342988] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] c01a5d24-eb46-4a69-993e-753880ce8e85/c01a5d24-eb46-4a69-993e-753880ce8e85.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 804.344759] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e335bff8-972a-4c35-9d2c-1e640ca84502 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.373306] env[63372]: DEBUG oslo_vmware.api [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024119, 'name': ReconfigVM_Task, 'duration_secs': 5.899614} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.375275] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.375663] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Reconfigured VM to detach interface {{(pid=63372) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 804.377861] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 804.377861] env[63372]: value = "task-1024137" [ 804.377861] env[63372]: _type = "Task" [ 804.377861] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.390749] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024137, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.510724] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.674876] env[63372]: DEBUG nova.network.neutron [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Updated VIF entry in instance network info cache for port bf61306f-1264-4480-b485-299826369f7d. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 804.675459] env[63372]: DEBUG nova.network.neutron [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Updating instance_info_cache with network_info: [{"id": "bf61306f-1264-4480-b485-299826369f7d", "address": "fa:16:3e:13:43:f9", "network": {"id": "2f241a45-cc18-46f6-b0a3-1aad8d2713fd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1953733697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a14a1aa8a0574b65a312754ae1091f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a88b707-352e-4be7-b1d6-ad6074b40ed9", "external-id": "nsx-vlan-transportzone-789", "segmentation_id": 789, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf61306f-12", "ovs_interfaceid": "bf61306f-1264-4480-b485-299826369f7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.701146] env[63372]: DEBUG oslo_concurrency.lockutils [None req-261f9a6b-b9e4-41e3-b3de-d369ef8cf6d2 tempest-ServerMetadataNegativeTestJSON-2093395660 tempest-ServerMetadataNegativeTestJSON-2093395660-project-member] Lock "4c9eb955-3bed-4b26-866f-b2a876b835dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.302s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.810698] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024135, 'name': CreateVM_Task, 'duration_secs': 0.626476} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.815163] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 804.818988] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.819399] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.820106] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 804.821618] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc471ec4-d9a9-4082-ba31-820fe0352189 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.831132] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 804.831132] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52a72218-4a56-2d5c-106c-68ef6ccfc2c8" [ 804.831132] env[63372]: _type = "Task" [ 804.831132] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.849624] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a72218-4a56-2d5c-106c-68ef6ccfc2c8, 'name': SearchDatastore_Task, 'duration_secs': 0.014905} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.857952] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.858083] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 804.858317] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.858480] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.858674] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 804.859587] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7b7d011-1f8c-456e-bb83-3813954d7cc4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.867847] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f01bf66c-2ebf-45c4-8f21-037d98ad148d tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "99f901a6-9bb3-4403-af0c-c8900f655cb3" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.868074] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f01bf66c-2ebf-45c4-8f21-037d98ad148d tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "99f901a6-9bb3-4403-af0c-c8900f655cb3" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.868308] env[63372]: DEBUG nova.compute.manager [None req-f01bf66c-2ebf-45c4-8f21-037d98ad148d tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 804.869364] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c4b56d-7a7d-4316-8be7-f5326e63238b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.874291] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 804.874608] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 804.880028] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93fedc75-7093-4593-a2ef-c96eca47d763 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.885192] env[63372]: DEBUG nova.compute.manager [req-66d1a9d3-33d3-4035-b661-11e6dfdf22df req-91f44d65-4b52-4cac-b61c-97a172e5d79b service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Received event network-vif-deleted-cd1ce4d9-46e7-41b8-8009-71283aa762b2 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.885436] env[63372]: INFO nova.compute.manager [req-66d1a9d3-33d3-4035-b661-11e6dfdf22df req-91f44d65-4b52-4cac-b61c-97a172e5d79b service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Neutron deleted interface cd1ce4d9-46e7-41b8-8009-71283aa762b2; detaching it from the instance and deleting it from the info cache [ 804.885720] env[63372]: DEBUG nova.network.neutron [req-66d1a9d3-33d3-4035-b661-11e6dfdf22df req-91f44d65-4b52-4cac-b61c-97a172e5d79b service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Updating instance_info_cache with network_info: [{"id": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "address": "fa:16:3e:17:c8:2e", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4cf27a3-fc", "ovs_interfaceid": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.892921] env[63372]: DEBUG nova.compute.manager [None req-f01bf66c-2ebf-45c4-8f21-037d98ad148d tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63372) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 804.895044] env[63372]: DEBUG nova.objects.instance [None req-f01bf66c-2ebf-45c4-8f21-037d98ad148d tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lazy-loading 'flavor' on Instance uuid 99f901a6-9bb3-4403-af0c-c8900f655cb3 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 804.901744] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 804.901744] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f3fdcd-5483-b0b9-bd0a-12a8e5657151" [ 804.901744] env[63372]: _type = "Task" [ 804.901744] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.908683] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024137, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.929144] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f3fdcd-5483-b0b9-bd0a-12a8e5657151, 'name': SearchDatastore_Task, 'duration_secs': 0.019309} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.931159] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0bc95f1-a235-4136-a0b6-01f04860120e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.939582] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 804.939582] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b3cc2a-6778-4322-03a3-c4f7f431dda3" [ 804.939582] env[63372]: _type = "Task" [ 804.939582] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.951630] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b3cc2a-6778-4322-03a3-c4f7f431dda3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.058353] env[63372]: DEBUG nova.network.neutron [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Successfully updated port: c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 805.180881] env[63372]: DEBUG oslo_concurrency.lockutils [req-50439025-4373-4330-983c-3b6559878fc8 req-5fddb949-bb19-4d4b-b72a-73db45cad5fd service nova] Releasing lock "refresh_cache-357505d0-f306-4e11-8a62-e03cfab2b7c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.339715] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1bb76a1-a770-48c7-90b1-20cc8cd477c4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.348367] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a86341-3d4e-4c46-afec-c358a528d474 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.383522] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c78f0b7-840a-4664-baf0-df2b9851b989 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.402156] env[63372]: DEBUG oslo_concurrency.lockutils [req-66d1a9d3-33d3-4035-b661-11e6dfdf22df req-91f44d65-4b52-4cac-b61c-97a172e5d79b service nova] Acquiring lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.402410] env[63372]: DEBUG oslo_concurrency.lockutils [req-66d1a9d3-33d3-4035-b661-11e6dfdf22df req-91f44d65-4b52-4cac-b61c-97a172e5d79b service nova] Acquired lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.404542] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024137, 'name': ReconfigVM_Task, 'duration_secs': 0.603275} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.405123] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f01bf66c-2ebf-45c4-8f21-037d98ad148d tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 805.405862] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-469e8507-e62f-4f08-88f9-3b506a477584 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.408560] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Reconfigured VM instance instance-00000041 to attach disk [datastore2] c01a5d24-eb46-4a69-993e-753880ce8e85/c01a5d24-eb46-4a69-993e-753880ce8e85.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 805.410203] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3e0e9a-6be2-45b0-bf19-9439abf65c4c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.414945] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e85c3bf6-0789-4465-b7ad-627d736d8cac {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.418576] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65287370-1dc6-497a-bd80-a2d19d56ad94 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.438059] env[63372]: DEBUG oslo_concurrency.lockutils [req-66d1a9d3-33d3-4035-b661-11e6dfdf22df req-91f44d65-4b52-4cac-b61c-97a172e5d79b service nova] Releasing lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.438059] env[63372]: WARNING nova.compute.manager [req-66d1a9d3-33d3-4035-b661-11e6dfdf22df req-91f44d65-4b52-4cac-b61c-97a172e5d79b service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Detach interface failed, port_id=cd1ce4d9-46e7-41b8-8009-71283aa762b2, reason: No device with interface-id cd1ce4d9-46e7-41b8-8009-71283aa762b2 exists on VM: nova.exception.NotFound: No device with interface-id cd1ce4d9-46e7-41b8-8009-71283aa762b2 exists on VM [ 805.448316] env[63372]: DEBUG nova.compute.provider_tree [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.449999] env[63372]: DEBUG oslo_vmware.api [None req-f01bf66c-2ebf-45c4-8f21-037d98ad148d tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 805.449999] env[63372]: value = "task-1024138" [ 805.449999] env[63372]: _type = "Task" [ 805.449999] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.450287] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 805.450287] env[63372]: value = "task-1024139" [ 805.450287] env[63372]: _type = "Task" [ 805.450287] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.472019] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b3cc2a-6778-4322-03a3-c4f7f431dda3, 'name': SearchDatastore_Task, 'duration_secs': 0.019725} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.479486] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.482824] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 357505d0-f306-4e11-8a62-e03cfab2b7c5/357505d0-f306-4e11-8a62-e03cfab2b7c5.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 805.482824] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024139, 'name': Rename_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.482824] env[63372]: DEBUG oslo_vmware.api [None req-f01bf66c-2ebf-45c4-8f21-037d98ad148d tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024138, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.482824] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e6c7bd99-121e-44a4-9c4e-51b3763b1db3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.489751] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 805.489751] env[63372]: value = "task-1024140" [ 805.489751] env[63372]: _type = "Task" [ 805.489751] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.499622] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024140, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.561032] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "refresh_cache-368101ee-e4fd-4fe4-b7c8-75c6b210189f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.561231] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired lock "refresh_cache-368101ee-e4fd-4fe4-b7c8-75c6b210189f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.561378] env[63372]: DEBUG nova.network.neutron [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 805.785146] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "f6675697-1529-46be-b28a-398ff3060d18" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.785934] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "f6675697-1529-46be-b28a-398ff3060d18" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.785934] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "f6675697-1529-46be-b28a-398ff3060d18-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.786132] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "f6675697-1529-46be-b28a-398ff3060d18-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.786352] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "f6675697-1529-46be-b28a-398ff3060d18-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.789872] env[63372]: INFO nova.compute.manager [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Terminating instance [ 805.792970] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "refresh_cache-f6675697-1529-46be-b28a-398ff3060d18" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.793172] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquired lock "refresh_cache-f6675697-1529-46be-b28a-398ff3060d18" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.793356] env[63372]: DEBUG nova.network.neutron [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 805.814755] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.814755] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.814755] env[63372]: DEBUG nova.network.neutron [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 805.952660] env[63372]: DEBUG nova.scheduler.client.report [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.987438] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024139, 'name': Rename_Task, 'duration_secs': 0.237585} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.992263] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 805.993187] env[63372]: DEBUG oslo_vmware.api [None req-f01bf66c-2ebf-45c4-8f21-037d98ad148d tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024138, 'name': PowerOffVM_Task, 'duration_secs': 0.332021} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.993292] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-746fab3c-9a3c-4f55-b4a3-78313002941e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.999017] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f01bf66c-2ebf-45c4-8f21-037d98ad148d tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 805.999314] env[63372]: DEBUG nova.compute.manager [None req-f01bf66c-2ebf-45c4-8f21-037d98ad148d tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 806.000551] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0abb0b-2f6d-4f83-9643-d3492d312a19 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.012675] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024140, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.016632] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 806.016632] env[63372]: value = "task-1024141" [ 806.016632] env[63372]: _type = "Task" [ 806.016632] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.033150] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024141, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.156162] env[63372]: DEBUG nova.network.neutron [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.333294] env[63372]: DEBUG nova.network.neutron [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.463837] env[63372]: DEBUG nova.network.neutron [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.479445] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.321s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.481999] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.497s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.483930] env[63372]: INFO nova.compute.claims [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.498430] env[63372]: DEBUG nova.network.neutron [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Updating instance_info_cache with network_info: [{"id": "c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa", "address": "fa:16:3e:03:9c:c3", "network": {"id": "7157ae73-79e9-454e-8673-7841f4cfe1f5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-786035586-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4548b52cd704cc0b054e2f7d1562b9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2cf91ff-12", "ovs_interfaceid": "c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.506915] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024140, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.675544} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.507269] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 357505d0-f306-4e11-8a62-e03cfab2b7c5/357505d0-f306-4e11-8a62-e03cfab2b7c5.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 806.507557] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 806.508251] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cdf0ef9b-beb8-468f-9d33-ea953c70b0f8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.511007] env[63372]: INFO nova.scheduler.client.report [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Deleted allocations for instance febeca1c-3370-4c74-8bd9-efad313df1e2 [ 806.518551] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 806.518551] env[63372]: value = "task-1024142" [ 806.518551] env[63372]: _type = "Task" [ 806.518551] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.530192] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f01bf66c-2ebf-45c4-8f21-037d98ad148d tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "99f901a6-9bb3-4403-af0c-c8900f655cb3" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.662s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.536653] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024142, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.541070] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024141, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.670461] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.670833] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.670976] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "bbba9f28-045d-41ab-8539-5b2968fe3d54-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.671129] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "bbba9f28-045d-41ab-8539-5b2968fe3d54-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.673853] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "bbba9f28-045d-41ab-8539-5b2968fe3d54-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.673853] env[63372]: INFO nova.compute.manager [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Terminating instance [ 806.676234] env[63372]: DEBUG nova.compute.manager [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 806.676423] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 806.677426] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43e26ff9-70db-4033-ae1f-406369b42d9b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.686766] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 806.687255] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-68b5b155-9b28-43b1-b374-61899fcb1d7a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.695830] env[63372]: DEBUG oslo_vmware.api [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 806.695830] env[63372]: value = "task-1024143" [ 806.695830] env[63372]: _type = "Task" [ 806.695830] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.707563] env[63372]: DEBUG oslo_vmware.api [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024143, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.968214] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Releasing lock "refresh_cache-f6675697-1529-46be-b28a-398ff3060d18" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.969018] env[63372]: DEBUG nova.compute.manager [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 806.969344] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 806.970602] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1c9750-df4b-42d3-b522-504a66820b1a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.983398] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 806.983713] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-846e1b9a-cfd1-47ea-b23e-5892000435da {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.996121] env[63372]: DEBUG oslo_vmware.api [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 806.996121] env[63372]: value = "task-1024144" [ 806.996121] env[63372]: _type = "Task" [ 806.996121] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.003737] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Releasing lock "refresh_cache-368101ee-e4fd-4fe4-b7c8-75c6b210189f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.003887] env[63372]: DEBUG nova.compute.manager [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Instance network_info: |[{"id": "c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa", "address": "fa:16:3e:03:9c:c3", "network": {"id": "7157ae73-79e9-454e-8673-7841f4cfe1f5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-786035586-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4548b52cd704cc0b054e2f7d1562b9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2cf91ff-12", "ovs_interfaceid": "c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 807.004343] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:9c:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.013885] env[63372]: DEBUG oslo.service.loopingcall [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.016429] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 807.025284] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-547e9862-48e4-47da-b63d-20e6767f0342 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.045477] env[63372]: DEBUG oslo_vmware.api [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024144, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.045477] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85ad56b1-efcb-4eb7-afb0-9a796b1fd72b tempest-ServersTestMultiNic-419105314 tempest-ServersTestMultiNic-419105314-project-member] Lock "febeca1c-3370-4c74-8bd9-efad313df1e2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.897s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.073718] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024142, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.247448} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.085665] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 807.086058] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.086058] env[63372]: value = "task-1024145" [ 807.086058] env[63372]: _type = "Task" [ 807.086058] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.086428] env[63372]: DEBUG oslo_vmware.api [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024141, 'name': PowerOnVM_Task, 'duration_secs': 1.04128} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.087964] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae7aad0-615d-4896-8225-2b291b5db9ec {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.093148] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 807.093716] env[63372]: INFO nova.compute.manager [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Took 9.79 seconds to spawn the instance on the hypervisor. [ 807.094116] env[63372]: DEBUG nova.compute.manager [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 807.101882] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c25b9a-744c-4139-bff1-bda716f20631 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.114202] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024145, 'name': CreateVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.136941] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 357505d0-f306-4e11-8a62-e03cfab2b7c5/357505d0-f306-4e11-8a62-e03cfab2b7c5.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 807.140464] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e74e1b1e-08e3-427b-8888-a94c78e914c0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.165083] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 807.165083] env[63372]: value = "task-1024146" [ 807.165083] env[63372]: _type = "Task" [ 807.165083] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.177915] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024146, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.199205] env[63372]: DEBUG nova.network.neutron [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Updating instance_info_cache with network_info: [{"id": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "address": "fa:16:3e:17:c8:2e", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4cf27a3-fc", "ovs_interfaceid": "b4cf27a3-fc17-4ac5-b96a-3be058f71e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.221674] env[63372]: DEBUG oslo_vmware.api [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024143, 'name': PowerOffVM_Task, 'duration_secs': 0.253852} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.222096] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 807.222317] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 807.222627] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-425f5e7b-45f3-4878-a664-51ab3f0968ab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.313195] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 807.313494] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 807.313787] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Deleting the datastore file [datastore2] bbba9f28-045d-41ab-8539-5b2968fe3d54 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 807.314966] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-586cc5d7-9c1d-48db-8e83-1d4e834a3ec4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.323187] env[63372]: DEBUG oslo_vmware.api [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 807.323187] env[63372]: value = "task-1024148" [ 807.323187] env[63372]: _type = "Task" [ 807.323187] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.332923] env[63372]: DEBUG oslo_vmware.api [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024148, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.439375] env[63372]: DEBUG nova.compute.manager [req-6d632929-a238-4c52-a0f2-2a4a369784e2 req-66d6fded-a87f-4228-9648-f6a223c547f9 service nova] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Received event network-vif-plugged-c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 807.439375] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d632929-a238-4c52-a0f2-2a4a369784e2 req-66d6fded-a87f-4228-9648-f6a223c547f9 service nova] Acquiring lock "368101ee-e4fd-4fe4-b7c8-75c6b210189f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.439375] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d632929-a238-4c52-a0f2-2a4a369784e2 req-66d6fded-a87f-4228-9648-f6a223c547f9 service nova] Lock "368101ee-e4fd-4fe4-b7c8-75c6b210189f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.439595] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d632929-a238-4c52-a0f2-2a4a369784e2 req-66d6fded-a87f-4228-9648-f6a223c547f9 service nova] Lock "368101ee-e4fd-4fe4-b7c8-75c6b210189f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.439847] env[63372]: DEBUG nova.compute.manager [req-6d632929-a238-4c52-a0f2-2a4a369784e2 req-66d6fded-a87f-4228-9648-f6a223c547f9 service nova] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] No waiting events found dispatching network-vif-plugged-c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 807.440043] env[63372]: WARNING nova.compute.manager [req-6d632929-a238-4c52-a0f2-2a4a369784e2 req-66d6fded-a87f-4228-9648-f6a223c547f9 service nova] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Received unexpected event network-vif-plugged-c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa for instance with vm_state building and task_state spawning. [ 807.440219] env[63372]: DEBUG nova.compute.manager [req-6d632929-a238-4c52-a0f2-2a4a369784e2 req-66d6fded-a87f-4228-9648-f6a223c547f9 service nova] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Received event network-changed-c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 807.440379] env[63372]: DEBUG nova.compute.manager [req-6d632929-a238-4c52-a0f2-2a4a369784e2 req-66d6fded-a87f-4228-9648-f6a223c547f9 service nova] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Refreshing instance network info cache due to event network-changed-c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 807.441195] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d632929-a238-4c52-a0f2-2a4a369784e2 req-66d6fded-a87f-4228-9648-f6a223c547f9 service nova] Acquiring lock "refresh_cache-368101ee-e4fd-4fe4-b7c8-75c6b210189f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.441195] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d632929-a238-4c52-a0f2-2a4a369784e2 req-66d6fded-a87f-4228-9648-f6a223c547f9 service nova] Acquired lock "refresh_cache-368101ee-e4fd-4fe4-b7c8-75c6b210189f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.441195] env[63372]: DEBUG nova.network.neutron [req-6d632929-a238-4c52-a0f2-2a4a369784e2 req-66d6fded-a87f-4228-9648-f6a223c547f9 service nova] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Refreshing network info cache for port c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 807.476000] env[63372]: DEBUG nova.objects.instance [None req-40c67cf4-77ee-4943-bcdb-94a18824f887 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Lazy-loading 'flavor' on Instance uuid d50919ac-3a0b-46ac-a837-ca3e6ad05173 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 807.511771] env[63372]: DEBUG oslo_vmware.api [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024144, 'name': PowerOffVM_Task, 'duration_secs': 0.296545} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.511771] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 807.512253] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 807.512933] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c765a13-4928-4d10-9564-22132ed69d07 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.555594] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 807.555884] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 807.556233] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Deleting the datastore file [datastore2] f6675697-1529-46be-b28a-398ff3060d18 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 807.556816] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-56e99082-1472-4409-951a-c52ae1c26b3b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.565572] env[63372]: DEBUG oslo_vmware.api [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 807.565572] env[63372]: value = "task-1024150" [ 807.565572] env[63372]: _type = "Task" [ 807.565572] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.578574] env[63372]: DEBUG oslo_vmware.api [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.608657] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024145, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.665201] env[63372]: INFO nova.compute.manager [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Took 30.44 seconds to build instance. [ 807.685704] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024146, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.712070] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "refresh_cache-bbba9f28-045d-41ab-8539-5b2968fe3d54" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.841580] env[63372]: DEBUG oslo_vmware.api [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024148, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.356369} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.842113] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 807.842311] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 807.842523] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 807.842708] env[63372]: INFO nova.compute.manager [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Took 1.17 seconds to destroy the instance on the hypervisor. [ 807.842977] env[63372]: DEBUG oslo.service.loopingcall [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.843190] env[63372]: DEBUG nova.compute.manager [-] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 807.843992] env[63372]: DEBUG nova.network.neutron [-] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 807.982990] env[63372]: DEBUG oslo_concurrency.lockutils [None req-40c67cf4-77ee-4943-bcdb-94a18824f887 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquiring lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.983239] env[63372]: DEBUG oslo_concurrency.lockutils [None req-40c67cf4-77ee-4943-bcdb-94a18824f887 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquired lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.081104] env[63372]: DEBUG oslo_vmware.api [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024150, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254999} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.081104] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 808.081104] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 808.081104] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 808.081104] env[63372]: INFO nova.compute.manager [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Took 1.11 seconds to destroy the instance on the hypervisor. [ 808.081104] env[63372]: DEBUG oslo.service.loopingcall [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.081104] env[63372]: DEBUG nova.compute.manager [-] [instance: f6675697-1529-46be-b28a-398ff3060d18] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.081104] env[63372]: DEBUG nova.network.neutron [-] [instance: f6675697-1529-46be-b28a-398ff3060d18] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 808.092987] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5752a260-a4a2-4904-96c1-525b112453f2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.107709] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942d460d-c49c-48dd-bb8f-09d50c886e54 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.114814] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024145, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.147888] env[63372]: DEBUG nova.network.neutron [-] [instance: f6675697-1529-46be-b28a-398ff3060d18] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 808.150373] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a46966-cf32-4447-a3d1-1581876d4fa3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.164494] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e703b5d0-2d48-4df5-9214-b56046a05cec {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.170981] env[63372]: DEBUG oslo_concurrency.lockutils [None req-802640c9-599a-4915-af46-74d8e90f8cbb tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.237s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.180780] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024146, 'name': ReconfigVM_Task, 'duration_secs': 0.608616} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.192875] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 357505d0-f306-4e11-8a62-e03cfab2b7c5/357505d0-f306-4e11-8a62-e03cfab2b7c5.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 808.193427] env[63372]: DEBUG nova.compute.provider_tree [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.195244] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-123ff505-eef9-485a-bba8-4cdd02bf4156 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.204062] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 808.204062] env[63372]: value = "task-1024151" [ 808.204062] env[63372]: _type = "Task" [ 808.204062] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.213999] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024151, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.219439] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1e7e88a8-738b-403d-99e3-e6137273a1d7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-bbba9f28-045d-41ab-8539-5b2968fe3d54-cd1ce4d9-46e7-41b8-8009-71283aa762b2" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.478s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.616126] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024145, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.650715] env[63372]: DEBUG nova.network.neutron [-] [instance: f6675697-1529-46be-b28a-398ff3060d18] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.659304] env[63372]: DEBUG nova.objects.instance [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lazy-loading 'flavor' on Instance uuid 99f901a6-9bb3-4403-af0c-c8900f655cb3 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 808.696615] env[63372]: DEBUG nova.scheduler.client.report [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.715685] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024151, 'name': Rename_Task, 'duration_secs': 0.244091} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.716908] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 808.717383] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4605ed9b-e1f9-4144-9815-4a0b0a5ded86 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.725862] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 808.725862] env[63372]: value = "task-1024152" [ 808.725862] env[63372]: _type = "Task" [ 808.725862] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.740760] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024152, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.963188] env[63372]: DEBUG nova.network.neutron [req-6d632929-a238-4c52-a0f2-2a4a369784e2 req-66d6fded-a87f-4228-9648-f6a223c547f9 service nova] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Updated VIF entry in instance network info cache for port c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 808.964986] env[63372]: DEBUG nova.network.neutron [req-6d632929-a238-4c52-a0f2-2a4a369784e2 req-66d6fded-a87f-4228-9648-f6a223c547f9 service nova] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Updating instance_info_cache with network_info: [{"id": "c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa", "address": "fa:16:3e:03:9c:c3", "network": {"id": "7157ae73-79e9-454e-8673-7841f4cfe1f5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-786035586-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4548b52cd704cc0b054e2f7d1562b9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2cf91ff-12", "ovs_interfaceid": "c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.122161] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024145, 'name': CreateVM_Task, 'duration_secs': 1.569575} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.122161] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 809.122926] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3c9be794-7dcb-43df-a9b2-d9289ac8889f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.123319] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3c9be794-7dcb-43df-a9b2-d9289ac8889f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.123982] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3c9be794-7dcb-43df-a9b2-d9289ac8889f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 809.124448] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8754146-000e-4a3d-afe5-a754d7ae101a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.134561] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 809.134561] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]528de25d-5fc3-b117-abbc-f546275e7dd3" [ 809.134561] env[63372]: _type = "Task" [ 809.134561] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.154870] env[63372]: INFO nova.compute.manager [-] [instance: f6675697-1529-46be-b28a-398ff3060d18] Took 1.07 seconds to deallocate network for instance. [ 809.157710] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3c9be794-7dcb-43df-a9b2-d9289ac8889f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.158277] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Processing image 3c9be794-7dcb-43df-a9b2-d9289ac8889f {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.158675] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3c9be794-7dcb-43df-a9b2-d9289ac8889f/3c9be794-7dcb-43df-a9b2-d9289ac8889f.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.159331] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3c9be794-7dcb-43df-a9b2-d9289ac8889f/3c9be794-7dcb-43df-a9b2-d9289ac8889f.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.159331] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.162284] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70193435-488a-4c1d-8ce0-076c6853c194 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.167170] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "refresh_cache-99f901a6-9bb3-4403-af0c-c8900f655cb3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.167489] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquired lock "refresh_cache-99f901a6-9bb3-4403-af0c-c8900f655cb3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.167731] env[63372]: DEBUG nova.network.neutron [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 809.167918] env[63372]: DEBUG nova.objects.instance [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lazy-loading 'info_cache' on Instance uuid 99f901a6-9bb3-4403-af0c-c8900f655cb3 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 809.190398] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.190653] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 809.191688] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6d6aa05-4e25-410a-b705-18e7bd27d45e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.202148] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 809.202148] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]524a95b8-feb3-58bd-4e46-0b955dda5a74" [ 809.202148] env[63372]: _type = "Task" [ 809.202148] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.208966] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.727s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.209713] env[63372]: DEBUG nova.compute.manager [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.213168] env[63372]: DEBUG oslo_concurrency.lockutils [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.617s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.213477] env[63372]: DEBUG nova.objects.instance [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Lazy-loading 'resources' on Instance uuid 7a75a51d-b804-453b-ba7b-fda6c4931802 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 809.227235] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524a95b8-feb3-58bd-4e46-0b955dda5a74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.239822] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024152, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.375648] env[63372]: DEBUG nova.network.neutron [None req-40c67cf4-77ee-4943-bcdb-94a18824f887 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 809.469019] env[63372]: DEBUG oslo_concurrency.lockutils [req-6d632929-a238-4c52-a0f2-2a4a369784e2 req-66d6fded-a87f-4228-9648-f6a223c547f9 service nova] Releasing lock "refresh_cache-368101ee-e4fd-4fe4-b7c8-75c6b210189f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.665152] env[63372]: DEBUG oslo_vmware.rw_handles [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ddbbba-ec2c-fc0d-54ae-3f9bb9b2fa1e/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 809.666329] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e28f1b-f651-4683-9412-c4e4d75e248e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.670109] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.672274] env[63372]: DEBUG nova.objects.base [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Object Instance<99f901a6-9bb3-4403-af0c-c8900f655cb3> lazy-loaded attributes: flavor,info_cache {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 809.674937] env[63372]: DEBUG oslo_vmware.rw_handles [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ddbbba-ec2c-fc0d-54ae-3f9bb9b2fa1e/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 809.675117] env[63372]: ERROR oslo_vmware.rw_handles [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ddbbba-ec2c-fc0d-54ae-3f9bb9b2fa1e/disk-0.vmdk due to incomplete transfer. [ 809.676150] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3c9e399a-862f-4029-95d3-56beba3e2192 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.688023] env[63372]: DEBUG oslo_vmware.rw_handles [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ddbbba-ec2c-fc0d-54ae-3f9bb9b2fa1e/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 809.688023] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Uploaded image de1edf71-70ee-4a4d-9486-cbfcce731c28 to the Glance image server {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 809.690993] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 809.694584] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-53de81ef-1ed8-4e1a-b664-859808a8359e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.702676] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 809.702676] env[63372]: value = "task-1024153" [ 809.702676] env[63372]: _type = "Task" [ 809.702676] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.720403] env[63372]: DEBUG nova.compute.utils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 809.722449] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024153, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.727767] env[63372]: DEBUG nova.compute.manager [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 809.732353] env[63372]: DEBUG nova.network.neutron [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 809.738969] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Preparing fetch location {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 809.739400] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Fetch image to [datastore2] OSTACK_IMG_81e4dbb4-984d-4a47-854d-c80457bfc995/OSTACK_IMG_81e4dbb4-984d-4a47-854d-c80457bfc995.vmdk {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 809.739695] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Downloading stream optimized image 3c9be794-7dcb-43df-a9b2-d9289ac8889f to [datastore2] OSTACK_IMG_81e4dbb4-984d-4a47-854d-c80457bfc995/OSTACK_IMG_81e4dbb4-984d-4a47-854d-c80457bfc995.vmdk on the data store datastore2 as vApp {{(pid=63372) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 809.739973] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Downloading image file data 3c9be794-7dcb-43df-a9b2-d9289ac8889f to the ESX as VM named 'OSTACK_IMG_81e4dbb4-984d-4a47-854d-c80457bfc995' {{(pid=63372) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 809.752259] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024152, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.793615] env[63372]: DEBUG nova.compute.manager [req-c0a577de-7b6a-4ee2-9a9c-2a198ec5b673 req-33aa1a54-c953-4c5b-bfa7-83acf6d9be84 service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Received event network-changed-c8f78ebe-604d-4414-a16d-ae25256f2730 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.793801] env[63372]: DEBUG nova.compute.manager [req-c0a577de-7b6a-4ee2-9a9c-2a198ec5b673 req-33aa1a54-c953-4c5b-bfa7-83acf6d9be84 service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Refreshing instance network info cache due to event network-changed-c8f78ebe-604d-4414-a16d-ae25256f2730. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 809.794000] env[63372]: DEBUG oslo_concurrency.lockutils [req-c0a577de-7b6a-4ee2-9a9c-2a198ec5b673 req-33aa1a54-c953-4c5b-bfa7-83acf6d9be84 service nova] Acquiring lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.837855] env[63372]: DEBUG nova.policy [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '881df5e623b340a487f37076c0cd4a15', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b6afb5579f2b48e4b094f6490136417f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 809.842897] env[63372]: DEBUG oslo_vmware.rw_handles [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 809.842897] env[63372]: value = "resgroup-9" [ 809.842897] env[63372]: _type = "ResourcePool" [ 809.842897] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 809.846347] env[63372]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-a2f9bfcc-ca2c-4df7-a254-7b9e7f6910fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.876714] env[63372]: DEBUG oslo_vmware.rw_handles [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lease: (returnval){ [ 809.876714] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5256b76b-7c15-9ce1-2ff3-f79a6352869c" [ 809.876714] env[63372]: _type = "HttpNfcLease" [ 809.876714] env[63372]: } obtained for vApp import into resource pool (val){ [ 809.876714] env[63372]: value = "resgroup-9" [ 809.876714] env[63372]: _type = "ResourcePool" [ 809.876714] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 809.876942] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the lease: (returnval){ [ 809.876942] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5256b76b-7c15-9ce1-2ff3-f79a6352869c" [ 809.876942] env[63372]: _type = "HttpNfcLease" [ 809.876942] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 809.890239] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 809.890239] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5256b76b-7c15-9ce1-2ff3-f79a6352869c" [ 809.890239] env[63372]: _type = "HttpNfcLease" [ 809.890239] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 809.998495] env[63372]: DEBUG nova.compute.manager [req-7c9bc772-7ab2-41c8-93fd-8e6e7c83369c req-903934ea-c557-4b1b-b15c-e79537c12594 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Received event network-changed-c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.998495] env[63372]: DEBUG nova.compute.manager [req-7c9bc772-7ab2-41c8-93fd-8e6e7c83369c req-903934ea-c557-4b1b-b15c-e79537c12594 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Refreshing instance network info cache due to event network-changed-c92e8cc9-dc72-4f20-b087-1d323c502108. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 809.998495] env[63372]: DEBUG oslo_concurrency.lockutils [req-7c9bc772-7ab2-41c8-93fd-8e6e7c83369c req-903934ea-c557-4b1b-b15c-e79537c12594 service nova] Acquiring lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.998495] env[63372]: DEBUG oslo_concurrency.lockutils [req-7c9bc772-7ab2-41c8-93fd-8e6e7c83369c req-903934ea-c557-4b1b-b15c-e79537c12594 service nova] Acquired lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.998495] env[63372]: DEBUG nova.network.neutron [req-7c9bc772-7ab2-41c8-93fd-8e6e7c83369c req-903934ea-c557-4b1b-b15c-e79537c12594 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Refreshing network info cache for port c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 810.226396] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024153, 'name': Destroy_Task, 'duration_secs': 0.42021} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.229215] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Destroyed the VM [ 810.229684] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 810.230416] env[63372]: DEBUG nova.compute.manager [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 810.236850] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-00506f33-7180-43d7-88d3-dc2234478955 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.251059] env[63372]: DEBUG nova.network.neutron [-] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.256987] env[63372]: DEBUG oslo_vmware.api [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024152, 'name': PowerOnVM_Task, 'duration_secs': 1.394983} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.257520] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 810.257520] env[63372]: value = "task-1024155" [ 810.257520] env[63372]: _type = "Task" [ 810.257520] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.258603] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 810.258603] env[63372]: INFO nova.compute.manager [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Took 9.00 seconds to spawn the instance on the hypervisor. [ 810.258752] env[63372]: DEBUG nova.compute.manager [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.259751] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d90a7a-80ec-4701-a527-dbf109055e83 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.278558] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024155, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.369740] env[63372]: DEBUG nova.network.neutron [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Successfully created port: 63d3f72b-cd36-4e27-bfac-e43665ed8ca1 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.375115] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f0add91-e6b1-48f3-9b96-457c8763ad82 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.389701] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 810.389701] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5256b76b-7c15-9ce1-2ff3-f79a6352869c" [ 810.389701] env[63372]: _type = "HttpNfcLease" [ 810.389701] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 810.393108] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06189230-0e37-4aea-a5e9-07207d3df4c1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.432561] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b5ba26-2b71-46b3-8073-5bc311a6b97b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.441184] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af886ec4-57d1-4d23-b76f-6f1cda4cdaf6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.456467] env[63372]: DEBUG nova.compute.provider_tree [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.605197] env[63372]: DEBUG nova.network.neutron [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Updating instance_info_cache with network_info: [{"id": "0e3060b6-6d21-4a2d-8e8a-ae48e64ee869", "address": "fa:16:3e:81:de:69", "network": {"id": "8171761d-bd68-4efd-ad4f-e661badde0c4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1048168720-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1cc87c31eb77441daeec0fad5bb1ce87", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e3060b6-6d", "ovs_interfaceid": "0e3060b6-6d21-4a2d-8e8a-ae48e64ee869", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.608294] env[63372]: DEBUG nova.network.neutron [None req-40c67cf4-77ee-4943-bcdb-94a18824f887 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Updating instance_info_cache with network_info: [{"id": "c8f78ebe-604d-4414-a16d-ae25256f2730", "address": "fa:16:3e:25:30:2b", "network": {"id": "73cbc616-354c-40de-971d-91fb5f0cbbac", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-817872371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2319173a3c445ca93ae2510141be32c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d", "external-id": "nsx-vlan-transportzone-591", "segmentation_id": 591, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8f78ebe-60", "ovs_interfaceid": "c8f78ebe-604d-4414-a16d-ae25256f2730", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.759681] env[63372]: INFO nova.compute.manager [-] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Took 2.92 seconds to deallocate network for instance. [ 810.775552] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024155, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.800579] env[63372]: INFO nova.compute.manager [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Took 33.16 seconds to build instance. [ 810.872353] env[63372]: DEBUG nova.network.neutron [req-7c9bc772-7ab2-41c8-93fd-8e6e7c83369c req-903934ea-c557-4b1b-b15c-e79537c12594 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updated VIF entry in instance network info cache for port c92e8cc9-dc72-4f20-b087-1d323c502108. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 810.872562] env[63372]: DEBUG nova.network.neutron [req-7c9bc772-7ab2-41c8-93fd-8e6e7c83369c req-903934ea-c557-4b1b-b15c-e79537c12594 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updating instance_info_cache with network_info: [{"id": "c92e8cc9-dc72-4f20-b087-1d323c502108", "address": "fa:16:3e:39:9e:be", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc92e8cc9-dc", "ovs_interfaceid": "c92e8cc9-dc72-4f20-b087-1d323c502108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.891445] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 810.891445] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5256b76b-7c15-9ce1-2ff3-f79a6352869c" [ 810.891445] env[63372]: _type = "HttpNfcLease" [ 810.891445] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 810.891757] env[63372]: DEBUG oslo_vmware.rw_handles [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 810.891757] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5256b76b-7c15-9ce1-2ff3-f79a6352869c" [ 810.891757] env[63372]: _type = "HttpNfcLease" [ 810.891757] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 810.892561] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e5fc6f-aaa7-405f-9f88-9855804ec113 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.901801] env[63372]: DEBUG oslo_vmware.rw_handles [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db191c-71ea-a5ca-1ac5-9efeb31b58ab/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 810.901996] env[63372]: DEBUG oslo_vmware.rw_handles [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db191c-71ea-a5ca-1ac5-9efeb31b58ab/disk-0.vmdk. {{(pid=63372) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 810.960757] env[63372]: DEBUG nova.scheduler.client.report [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 810.969564] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1d2b9a57-20fb-4153-8a2f-8b9216a57067 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.108509] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Releasing lock "refresh_cache-99f901a6-9bb3-4403-af0c-c8900f655cb3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.110389] env[63372]: DEBUG oslo_concurrency.lockutils [None req-40c67cf4-77ee-4943-bcdb-94a18824f887 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Releasing lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.110610] env[63372]: DEBUG nova.compute.manager [None req-40c67cf4-77ee-4943-bcdb-94a18824f887 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Inject network info {{(pid=63372) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 811.110863] env[63372]: DEBUG nova.compute.manager [None req-40c67cf4-77ee-4943-bcdb-94a18824f887 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] network_info to inject: |[{"id": "c8f78ebe-604d-4414-a16d-ae25256f2730", "address": "fa:16:3e:25:30:2b", "network": {"id": "73cbc616-354c-40de-971d-91fb5f0cbbac", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-817872371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2319173a3c445ca93ae2510141be32c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d", "external-id": "nsx-vlan-transportzone-591", "segmentation_id": 591, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8f78ebe-60", "ovs_interfaceid": "c8f78ebe-604d-4414-a16d-ae25256f2730", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 811.117016] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-40c67cf4-77ee-4943-bcdb-94a18824f887 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Reconfiguring VM instance to set the machine id {{(pid=63372) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 811.117972] env[63372]: DEBUG oslo_concurrency.lockutils [req-c0a577de-7b6a-4ee2-9a9c-2a198ec5b673 req-33aa1a54-c953-4c5b-bfa7-83acf6d9be84 service nova] Acquired lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.118229] env[63372]: DEBUG nova.network.neutron [req-c0a577de-7b6a-4ee2-9a9c-2a198ec5b673 req-33aa1a54-c953-4c5b-bfa7-83acf6d9be84 service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Refreshing network info cache for port c8f78ebe-604d-4414-a16d-ae25256f2730 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 811.119377] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d09ae1d-026c-4a37-b21a-c80999b2806f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.145649] env[63372]: DEBUG oslo_vmware.api [None req-40c67cf4-77ee-4943-bcdb-94a18824f887 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Waiting for the task: (returnval){ [ 811.145649] env[63372]: value = "task-1024156" [ 811.145649] env[63372]: _type = "Task" [ 811.145649] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.158111] env[63372]: DEBUG oslo_vmware.api [None req-40c67cf4-77ee-4943-bcdb-94a18824f887 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1024156, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.247641] env[63372]: DEBUG nova.compute.manager [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 811.269512] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.278771] env[63372]: DEBUG oslo_vmware.api [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024155, 'name': RemoveSnapshot_Task, 'duration_secs': 0.586271} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.283052] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 811.283262] env[63372]: INFO nova.compute.manager [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Took 16.80 seconds to snapshot the instance on the hypervisor. [ 811.299129] env[63372]: DEBUG nova.virt.hardware [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.299639] env[63372]: DEBUG nova.virt.hardware [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.299639] env[63372]: DEBUG nova.virt.hardware [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.299789] env[63372]: DEBUG nova.virt.hardware [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.299877] env[63372]: DEBUG nova.virt.hardware [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.300249] env[63372]: DEBUG nova.virt.hardware [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.300584] env[63372]: DEBUG nova.virt.hardware [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.300897] env[63372]: DEBUG nova.virt.hardware [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.301116] env[63372]: DEBUG nova.virt.hardware [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.301296] env[63372]: DEBUG nova.virt.hardware [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.301479] env[63372]: DEBUG nova.virt.hardware [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.303971] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81ac34f-e335-4e19-a9cc-6d103447e8e6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.312198] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6bf183c0-4c15-467c-8449-d711ba44d6c8 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "357505d0-f306-4e11-8a62-e03cfab2b7c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.727s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.327270] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad661667-399f-44ea-a9d5-86fc4103080d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.378100] env[63372]: DEBUG oslo_concurrency.lockutils [req-7c9bc772-7ab2-41c8-93fd-8e6e7c83369c req-903934ea-c557-4b1b-b15c-e79537c12594 service nova] Releasing lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.447196] env[63372]: DEBUG nova.network.neutron [req-c0a577de-7b6a-4ee2-9a9c-2a198ec5b673 req-33aa1a54-c953-4c5b-bfa7-83acf6d9be84 service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Updated VIF entry in instance network info cache for port c8f78ebe-604d-4414-a16d-ae25256f2730. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 811.447896] env[63372]: DEBUG nova.network.neutron [req-c0a577de-7b6a-4ee2-9a9c-2a198ec5b673 req-33aa1a54-c953-4c5b-bfa7-83acf6d9be84 service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Updating instance_info_cache with network_info: [{"id": "c8f78ebe-604d-4414-a16d-ae25256f2730", "address": "fa:16:3e:25:30:2b", "network": {"id": "73cbc616-354c-40de-971d-91fb5f0cbbac", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-817872371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2319173a3c445ca93ae2510141be32c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d", "external-id": "nsx-vlan-transportzone-591", "segmentation_id": 591, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8f78ebe-60", "ovs_interfaceid": "c8f78ebe-604d-4414-a16d-ae25256f2730", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.467767] env[63372]: DEBUG oslo_concurrency.lockutils [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.254s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.475683] env[63372]: DEBUG oslo_concurrency.lockutils [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 20.716s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.518467] env[63372]: DEBUG nova.objects.instance [None req-85480171-5359-4a60-93fe-bf5a0e428c51 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Lazy-loading 'flavor' on Instance uuid d50919ac-3a0b-46ac-a837-ca3e6ad05173 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 811.519691] env[63372]: INFO nova.scheduler.client.report [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Deleted allocations for instance 7a75a51d-b804-453b-ba7b-fda6c4931802 [ 811.537779] env[63372]: INFO nova.compute.manager [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Rescuing [ 811.538136] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "refresh_cache-357505d0-f306-4e11-8a62-e03cfab2b7c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.538344] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquired lock "refresh_cache-357505d0-f306-4e11-8a62-e03cfab2b7c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.538545] env[63372]: DEBUG nova.network.neutron [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 811.619375] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 811.619781] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c1e2a80-b735-4270-9432-e3c06af10b0b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.630313] env[63372]: DEBUG oslo_vmware.api [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 811.630313] env[63372]: value = "task-1024157" [ 811.630313] env[63372]: _type = "Task" [ 811.630313] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.641158] env[63372]: DEBUG oslo_vmware.api [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024157, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.658615] env[63372]: DEBUG oslo_vmware.api [None req-40c67cf4-77ee-4943-bcdb-94a18824f887 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1024156, 'name': ReconfigVM_Task, 'duration_secs': 0.212763} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.658970] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-40c67cf4-77ee-4943-bcdb-94a18824f887 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Reconfigured VM instance to set the machine id {{(pid=63372) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 811.723414] env[63372]: DEBUG oslo_vmware.rw_handles [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Completed reading data from the image iterator. {{(pid=63372) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 811.723654] env[63372]: DEBUG oslo_vmware.rw_handles [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db191c-71ea-a5ca-1ac5-9efeb31b58ab/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 811.724600] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86dd8351-2c3a-42dd-abdf-879d38c616ee {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.733435] env[63372]: DEBUG oslo_vmware.rw_handles [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db191c-71ea-a5ca-1ac5-9efeb31b58ab/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 811.733590] env[63372]: DEBUG oslo_vmware.rw_handles [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db191c-71ea-a5ca-1ac5-9efeb31b58ab/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 811.733845] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-2d63ced1-840a-4fda-bc9f-bbbe2e50b410 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.862827] env[63372]: DEBUG nova.compute.manager [None req-2146d250-0f2c-42aa-891a-e4cebb6d3d34 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Found 1 images (rotation: 2) {{(pid=63372) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 811.951159] env[63372]: DEBUG oslo_concurrency.lockutils [req-c0a577de-7b6a-4ee2-9a9c-2a198ec5b673 req-33aa1a54-c953-4c5b-bfa7-83acf6d9be84 service nova] Releasing lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.978677] env[63372]: DEBUG nova.objects.instance [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lazy-loading 'migration_context' on Instance uuid 8283b736-ad02-4082-97b7-561bd5c5da93 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.025665] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85480171-5359-4a60-93fe-bf5a0e428c51 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquiring lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.025861] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85480171-5359-4a60-93fe-bf5a0e428c51 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquired lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.032778] env[63372]: DEBUG oslo_concurrency.lockutils [None req-982ce035-a664-4132-ba5d-d52303785621 tempest-ServersV294TestFqdnHostnames-532803069 tempest-ServersV294TestFqdnHostnames-532803069-project-member] Lock "7a75a51d-b804-453b-ba7b-fda6c4931802" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.893s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.063886] env[63372]: DEBUG oslo_vmware.rw_handles [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db191c-71ea-a5ca-1ac5-9efeb31b58ab/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 812.064192] env[63372]: INFO nova.virt.vmwareapi.images [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Downloaded image file data 3c9be794-7dcb-43df-a9b2-d9289ac8889f [ 812.065307] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bccda6de-a1eb-4823-a6ab-e2b1712e9291 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.072117] env[63372]: DEBUG nova.compute.manager [req-99eb41c7-6f11-449e-b0b0-e6cd8931d024 req-94471d7d-4fcf-4ba1-b00d-c94cafea1bac service nova] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Received event network-vif-deleted-b4cf27a3-fc17-4ac5-b96a-3be058f71e03 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.087517] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fa51bcb2-0854-4b15-8520-3f459ccc72c2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.146478] env[63372]: DEBUG oslo_vmware.api [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024157, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.651549] env[63372]: DEBUG oslo_vmware.api [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024157, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.726904] env[63372]: DEBUG nova.compute.manager [req-fac0b0bf-a135-40e5-8c82-e8a75ad54d33 req-63b25593-e64d-4a02-9abe-5ca3968dee34 service nova] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Received event network-vif-plugged-63d3f72b-cd36-4e27-bfac-e43665ed8ca1 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.727031] env[63372]: DEBUG oslo_concurrency.lockutils [req-fac0b0bf-a135-40e5-8c82-e8a75ad54d33 req-63b25593-e64d-4a02-9abe-5ca3968dee34 service nova] Acquiring lock "89301344-84a9-4d13-aae7-99943d0a478e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.727255] env[63372]: DEBUG oslo_concurrency.lockutils [req-fac0b0bf-a135-40e5-8c82-e8a75ad54d33 req-63b25593-e64d-4a02-9abe-5ca3968dee34 service nova] Lock "89301344-84a9-4d13-aae7-99943d0a478e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.727419] env[63372]: DEBUG oslo_concurrency.lockutils [req-fac0b0bf-a135-40e5-8c82-e8a75ad54d33 req-63b25593-e64d-4a02-9abe-5ca3968dee34 service nova] Lock "89301344-84a9-4d13-aae7-99943d0a478e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.727582] env[63372]: DEBUG nova.compute.manager [req-fac0b0bf-a135-40e5-8c82-e8a75ad54d33 req-63b25593-e64d-4a02-9abe-5ca3968dee34 service nova] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] No waiting events found dispatching network-vif-plugged-63d3f72b-cd36-4e27-bfac-e43665ed8ca1 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 812.729704] env[63372]: WARNING nova.compute.manager [req-fac0b0bf-a135-40e5-8c82-e8a75ad54d33 req-63b25593-e64d-4a02-9abe-5ca3968dee34 service nova] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Received unexpected event network-vif-plugged-63d3f72b-cd36-4e27-bfac-e43665ed8ca1 for instance with vm_state building and task_state spawning. [ 812.905810] env[63372]: DEBUG nova.network.neutron [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Updating instance_info_cache with network_info: [{"id": "bf61306f-1264-4480-b485-299826369f7d", "address": "fa:16:3e:13:43:f9", "network": {"id": "2f241a45-cc18-46f6-b0a3-1aad8d2713fd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1953733697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a14a1aa8a0574b65a312754ae1091f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a88b707-352e-4be7-b1d6-ad6074b40ed9", "external-id": "nsx-vlan-transportzone-789", "segmentation_id": 789, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf61306f-12", "ovs_interfaceid": "bf61306f-1264-4480-b485-299826369f7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.919467] env[63372]: DEBUG nova.network.neutron [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Successfully updated port: 63d3f72b-cd36-4e27-bfac-e43665ed8ca1 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 812.988566] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3244f15-a8bf-4332-9e11-d2f5692bb4bf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.998355] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac17125-6e99-480c-8319-9223e4f56757 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.005951] env[63372]: DEBUG nova.network.neutron [None req-85480171-5359-4a60-93fe-bf5a0e428c51 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 813.039489] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27809370-db34-4b3c-a65b-12b5abcecb57 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.048803] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-416bd0b1-5367-446b-9b2e-d2f8ae7f2803 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.064160] env[63372]: DEBUG nova.compute.provider_tree [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.095683] env[63372]: DEBUG nova.compute.manager [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 813.096758] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c1315c-18a4-4b02-b657-6d0a38a7d8c6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.147020] env[63372]: DEBUG oslo_vmware.api [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024157, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.408584] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Releasing lock "refresh_cache-357505d0-f306-4e11-8a62-e03cfab2b7c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.422580] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "refresh_cache-89301344-84a9-4d13-aae7-99943d0a478e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.422828] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquired lock "refresh_cache-89301344-84a9-4d13-aae7-99943d0a478e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.422904] env[63372]: DEBUG nova.network.neutron [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 813.568282] env[63372]: DEBUG nova.scheduler.client.report [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.609405] env[63372]: INFO nova.compute.manager [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] instance snapshotting [ 813.610410] env[63372]: DEBUG nova.objects.instance [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'flavor' on Instance uuid 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 813.646426] env[63372]: DEBUG oslo_vmware.api [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024157, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.741991] env[63372]: INFO nova.virt.vmwareapi.images [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] The imported VM was unregistered [ 813.745302] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Caching image {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 813.745679] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Creating directory with path [datastore2] devstack-image-cache_base/3c9be794-7dcb-43df-a9b2-d9289ac8889f {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 813.748547] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9d76fa8a-2804-4a9c-877b-a092bd0824c9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.764598] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Created directory with path [datastore2] devstack-image-cache_base/3c9be794-7dcb-43df-a9b2-d9289ac8889f {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 813.765140] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_81e4dbb4-984d-4a47-854d-c80457bfc995/OSTACK_IMG_81e4dbb4-984d-4a47-854d-c80457bfc995.vmdk to [datastore2] devstack-image-cache_base/3c9be794-7dcb-43df-a9b2-d9289ac8889f/3c9be794-7dcb-43df-a9b2-d9289ac8889f.vmdk. {{(pid=63372) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 813.765140] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-052bfa6f-cac6-4709-9357-773720756210 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.774373] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 813.774373] env[63372]: value = "task-1024159" [ 813.774373] env[63372]: _type = "Task" [ 813.774373] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.783723] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024159, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.951506] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 813.951877] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-73894429-46c6-4e68-8ead-32dfa29b4061 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.962552] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 813.962552] env[63372]: value = "task-1024160" [ 813.962552] env[63372]: _type = "Task" [ 813.962552] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.973139] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024160, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.017927] env[63372]: DEBUG nova.network.neutron [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 814.130161] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85949439-592c-4bc8-bcc0-9c02a2dcb19e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.158389] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1750bd6b-d863-483f-8bbf-2fd29f745daf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.165366] env[63372]: DEBUG oslo_vmware.api [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024157, 'name': PowerOnVM_Task, 'duration_secs': 2.124724} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.166125] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 814.166353] env[63372]: DEBUG nova.compute.manager [None req-da76d78b-ad1f-445f-a97d-afba70914700 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.167781] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abfef4be-7da7-4f88-aac1-d69b507a9095 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.275929] env[63372]: DEBUG nova.network.neutron [None req-85480171-5359-4a60-93fe-bf5a0e428c51 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Updating instance_info_cache with network_info: [{"id": "c8f78ebe-604d-4414-a16d-ae25256f2730", "address": "fa:16:3e:25:30:2b", "network": {"id": "73cbc616-354c-40de-971d-91fb5f0cbbac", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-817872371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2319173a3c445ca93ae2510141be32c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d", "external-id": "nsx-vlan-transportzone-591", "segmentation_id": 591, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8f78ebe-60", "ovs_interfaceid": "c8f78ebe-604d-4414-a16d-ae25256f2730", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.291258] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024159, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.453446] env[63372]: DEBUG nova.network.neutron [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Updating instance_info_cache with network_info: [{"id": "63d3f72b-cd36-4e27-bfac-e43665ed8ca1", "address": "fa:16:3e:da:87:2c", "network": {"id": "b924481a-4e51-4483-a710-f01fc39c851c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1289474667-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b6afb5579f2b48e4b094f6490136417f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63d3f72b-cd", "ovs_interfaceid": "63d3f72b-cd36-4e27-bfac-e43665ed8ca1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.475216] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024160, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.580939] env[63372]: DEBUG oslo_concurrency.lockutils [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.105s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.595147] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.694s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.595471] env[63372]: DEBUG nova.objects.instance [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lazy-loading 'resources' on Instance uuid bad91d77-c7ee-4572-b1ed-068b2a55233c {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 814.679568] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 814.679936] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-07fd9cef-2e28-4281-a554-20025f0a2f12 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.691015] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 814.691015] env[63372]: value = "task-1024161" [ 814.691015] env[63372]: _type = "Task" [ 814.691015] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.706742] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024161, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.746976] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Acquiring lock "69a107d0-80c7-42e5-b514-b4273e1a3359" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.747248] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Lock "69a107d0-80c7-42e5-b514-b4273e1a3359" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.747739] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Acquiring lock "69a107d0-80c7-42e5-b514-b4273e1a3359-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.747739] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Lock "69a107d0-80c7-42e5-b514-b4273e1a3359-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.747878] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Lock "69a107d0-80c7-42e5-b514-b4273e1a3359-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.753560] env[63372]: INFO nova.compute.manager [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Terminating instance [ 814.757415] env[63372]: DEBUG nova.compute.manager [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 814.757781] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 814.759979] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e7f3e56-1347-4a02-af33-3238202710f3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.766601] env[63372]: DEBUG nova.compute.manager [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Received event network-changed-63d3f72b-cd36-4e27-bfac-e43665ed8ca1 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 814.766761] env[63372]: DEBUG nova.compute.manager [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Refreshing instance network info cache due to event network-changed-63d3f72b-cd36-4e27-bfac-e43665ed8ca1. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 814.767205] env[63372]: DEBUG oslo_concurrency.lockutils [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] Acquiring lock "refresh_cache-89301344-84a9-4d13-aae7-99943d0a478e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.776132] env[63372]: DEBUG oslo_vmware.api [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Waiting for the task: (returnval){ [ 814.776132] env[63372]: value = "task-1024162" [ 814.776132] env[63372]: _type = "Task" [ 814.776132] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.786813] env[63372]: DEBUG oslo_concurrency.lockutils [None req-85480171-5359-4a60-93fe-bf5a0e428c51 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Releasing lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.787102] env[63372]: DEBUG nova.compute.manager [None req-85480171-5359-4a60-93fe-bf5a0e428c51 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Inject network info {{(pid=63372) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 814.787415] env[63372]: DEBUG nova.compute.manager [None req-85480171-5359-4a60-93fe-bf5a0e428c51 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] network_info to inject: |[{"id": "c8f78ebe-604d-4414-a16d-ae25256f2730", "address": "fa:16:3e:25:30:2b", "network": {"id": "73cbc616-354c-40de-971d-91fb5f0cbbac", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-817872371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2319173a3c445ca93ae2510141be32c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d", "external-id": "nsx-vlan-transportzone-591", "segmentation_id": 591, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8f78ebe-60", "ovs_interfaceid": "c8f78ebe-604d-4414-a16d-ae25256f2730", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 814.792914] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-85480171-5359-4a60-93fe-bf5a0e428c51 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Reconfiguring VM instance to set the machine id {{(pid=63372) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 814.793957] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bbb2859-66c1-4010-8e1e-7a08836bce80 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.815347] env[63372]: DEBUG oslo_vmware.api [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1024162, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.818666] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024159, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.820560] env[63372]: DEBUG oslo_vmware.api [None req-85480171-5359-4a60-93fe-bf5a0e428c51 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Waiting for the task: (returnval){ [ 814.820560] env[63372]: value = "task-1024163" [ 814.820560] env[63372]: _type = "Task" [ 814.820560] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.830080] env[63372]: DEBUG oslo_vmware.api [None req-85480171-5359-4a60-93fe-bf5a0e428c51 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1024163, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.958378] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Releasing lock "refresh_cache-89301344-84a9-4d13-aae7-99943d0a478e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.959336] env[63372]: DEBUG nova.compute.manager [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Instance network_info: |[{"id": "63d3f72b-cd36-4e27-bfac-e43665ed8ca1", "address": "fa:16:3e:da:87:2c", "network": {"id": "b924481a-4e51-4483-a710-f01fc39c851c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1289474667-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b6afb5579f2b48e4b094f6490136417f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63d3f72b-cd", "ovs_interfaceid": "63d3f72b-cd36-4e27-bfac-e43665ed8ca1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 814.959838] env[63372]: DEBUG oslo_concurrency.lockutils [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] Acquired lock "refresh_cache-89301344-84a9-4d13-aae7-99943d0a478e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.960144] env[63372]: DEBUG nova.network.neutron [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Refreshing network info cache for port 63d3f72b-cd36-4e27-bfac-e43665ed8ca1 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 814.964326] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:87:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '40c947c4-f471-4d48-8e43-fee54198107e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '63d3f72b-cd36-4e27-bfac-e43665ed8ca1', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 814.983487] env[63372]: DEBUG oslo.service.loopingcall [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.989941] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 814.996609] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-086559d4-b496-45f6-b563-6d1a97661901 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.023671] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024160, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.025230] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 815.025230] env[63372]: value = "task-1024164" [ 815.025230] env[63372]: _type = "Task" [ 815.025230] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.041024] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024164, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.208689] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024161, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.305525] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024159, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.307466] env[63372]: DEBUG oslo_vmware.api [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1024162, 'name': PowerOffVM_Task, 'duration_secs': 0.242561} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.312312] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 815.312720] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Volume detach. Driver type: vmdk {{(pid=63372) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 815.312973] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227276', 'volume_id': '3226fd91-2732-41ce-bb8e-f909a830c8d4', 'name': 'volume-3226fd91-2732-41ce-bb8e-f909a830c8d4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '69a107d0-80c7-42e5-b514-b4273e1a3359', 'attached_at': '', 'detached_at': '', 'volume_id': '3226fd91-2732-41ce-bb8e-f909a830c8d4', 'serial': '3226fd91-2732-41ce-bb8e-f909a830c8d4'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 815.317023] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc47ca2d-9d44-436f-ab81-034d673a974f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.349927] env[63372]: DEBUG oslo_vmware.api [None req-85480171-5359-4a60-93fe-bf5a0e428c51 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1024163, 'name': ReconfigVM_Task, 'duration_secs': 0.242152} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.353235] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-85480171-5359-4a60-93fe-bf5a0e428c51 tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Reconfigured VM instance to set the machine id {{(pid=63372) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 815.356841] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6855b6b3-e42b-45ad-a41d-1b3d17fce8ba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.367108] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0fefc00-2de1-496d-ae33-129ae49e0105 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.371213] env[63372]: DEBUG nova.network.neutron [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Updated VIF entry in instance network info cache for port 63d3f72b-cd36-4e27-bfac-e43665ed8ca1. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 815.371213] env[63372]: DEBUG nova.network.neutron [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Updating instance_info_cache with network_info: [{"id": "63d3f72b-cd36-4e27-bfac-e43665ed8ca1", "address": "fa:16:3e:da:87:2c", "network": {"id": "b924481a-4e51-4483-a710-f01fc39c851c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1289474667-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b6afb5579f2b48e4b094f6490136417f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63d3f72b-cd", "ovs_interfaceid": "63d3f72b-cd36-4e27-bfac-e43665ed8ca1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.397759] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2268981-9497-4445-b83b-730e8599bc18 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.417796] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] The volume has not been displaced from its original location: [datastore2] volume-3226fd91-2732-41ce-bb8e-f909a830c8d4/volume-3226fd91-2732-41ce-bb8e-f909a830c8d4.vmdk. No consolidation needed. {{(pid=63372) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 815.424391] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Reconfiguring VM instance instance-00000037 to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 815.427382] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-921ba0af-a3dd-4738-b3bc-b34b1569773d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.462171] env[63372]: DEBUG oslo_vmware.api [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Waiting for the task: (returnval){ [ 815.462171] env[63372]: value = "task-1024165" [ 815.462171] env[63372]: _type = "Task" [ 815.462171] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.476455] env[63372]: DEBUG oslo_vmware.api [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1024165, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.504391] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024160, 'name': PowerOffVM_Task, 'duration_secs': 1.125689} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.507589] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 815.508769] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931b8e7a-ce9c-4206-b060-99ee8da8e09f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.535210] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8699e9-2992-46d0-9411-4812a7254d25 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.549617] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024164, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.585354] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 815.585712] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25330c20-eabd-43d5-9926-c2ae6ba94505 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.595065] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 815.595065] env[63372]: value = "task-1024166" [ 815.595065] env[63372]: _type = "Task" [ 815.595065] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.606814] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] VM already powered off {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 815.606814] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.607065] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.607222] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.607409] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.607729] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ce97acb-4b1b-468d-9036-5514cb5cbf38 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.631173] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.631173] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 815.637017] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de347820-8efe-4313-9db2-0aa2f7e9615b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.646339] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 815.646339] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52ea8439-2170-3922-71d0-a068b39f18ab" [ 815.646339] env[63372]: _type = "Task" [ 815.646339] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.656137] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ea8439-2170-3922-71d0-a068b39f18ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.710110] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024161, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.749796] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca468b0c-7088-46ef-8df7-2dcc92fc4990 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.759936] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5dc6f6-77af-4a88-be94-2a032295d83f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.796744] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3bdde2b-6bed-4598-ab28-d2d036252b2f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.808351] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024159, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.812079] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703f9b7d-3aaa-466d-bb4c-3883ec28fa07 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.832749] env[63372]: DEBUG nova.compute.provider_tree [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.873817] env[63372]: DEBUG oslo_concurrency.lockutils [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] Releasing lock "refresh_cache-89301344-84a9-4d13-aae7-99943d0a478e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.874127] env[63372]: DEBUG nova.compute.manager [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Received event network-changed-c8f78ebe-604d-4414-a16d-ae25256f2730 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 815.874303] env[63372]: DEBUG nova.compute.manager [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Refreshing instance network info cache due to event network-changed-c8f78ebe-604d-4414-a16d-ae25256f2730. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 815.874520] env[63372]: DEBUG oslo_concurrency.lockutils [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] Acquiring lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.874660] env[63372]: DEBUG oslo_concurrency.lockutils [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] Acquired lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.874823] env[63372]: DEBUG nova.network.neutron [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Refreshing network info cache for port c8f78ebe-604d-4414-a16d-ae25256f2730 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 815.973729] env[63372]: DEBUG oslo_vmware.api [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1024165, 'name': ReconfigVM_Task, 'duration_secs': 0.239043} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.974083] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Reconfigured VM instance instance-00000037 to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 815.981042] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f57c5f89-bf87-460f-a50a-dbfdb3d58868 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.998491] env[63372]: DEBUG oslo_vmware.api [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Waiting for the task: (returnval){ [ 815.998491] env[63372]: value = "task-1024167" [ 815.998491] env[63372]: _type = "Task" [ 815.998491] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.013028] env[63372]: DEBUG oslo_vmware.api [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1024167, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.041746] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024164, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.134013] env[63372]: INFO nova.compute.manager [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Swapping old allocation on dict_keys(['c1b110ca-a185-44c2-ba1c-f05b3b420add']) held by migration e4ca8d01-990e-4e1b-8d4a-067812994b8f for instance [ 816.160022] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ea8439-2170-3922-71d0-a068b39f18ab, 'name': SearchDatastore_Task, 'duration_secs': 0.084019} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.160960] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7fcdb31-556d-4a3b-bab4-3f3fba47ede7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.164375] env[63372]: DEBUG nova.scheduler.client.report [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Overwriting current allocation {'allocations': {'c1b110ca-a185-44c2-ba1c-f05b3b420add': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 89}}, 'project_id': '5cf13f5cc32c4b3ebaee47f03bb64ba7', 'user_id': '1048d032dc00441f9c2f99b60f357dfb', 'consumer_generation': 1} on consumer 8283b736-ad02-4082-97b7-561bd5c5da93 {{(pid=63372) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 816.170783] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 816.170783] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52bd7cde-ada3-55af-c9a0-eb02b57ea4b8" [ 816.170783] env[63372]: _type = "Task" [ 816.170783] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.174980] env[63372]: DEBUG oslo_concurrency.lockutils [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquiring lock "d50919ac-3a0b-46ac-a837-ca3e6ad05173" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.175274] env[63372]: DEBUG oslo_concurrency.lockutils [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Lock "d50919ac-3a0b-46ac-a837-ca3e6ad05173" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.175479] env[63372]: DEBUG oslo_concurrency.lockutils [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquiring lock "d50919ac-3a0b-46ac-a837-ca3e6ad05173-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.175662] env[63372]: DEBUG oslo_concurrency.lockutils [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Lock "d50919ac-3a0b-46ac-a837-ca3e6ad05173-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.175836] env[63372]: DEBUG oslo_concurrency.lockutils [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Lock "d50919ac-3a0b-46ac-a837-ca3e6ad05173-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.178580] env[63372]: INFO nova.compute.manager [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Terminating instance [ 816.185335] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52bd7cde-ada3-55af-c9a0-eb02b57ea4b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.186141] env[63372]: DEBUG nova.compute.manager [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 816.186405] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 816.187384] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799265e5-857a-498f-b156-bb4ce4d98c40 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.200867] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 816.201561] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-787ca6d0-fa29-4530-9e5e-811543b3aea7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.209956] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024161, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.211685] env[63372]: DEBUG oslo_vmware.api [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Waiting for the task: (returnval){ [ 816.211685] env[63372]: value = "task-1024168" [ 816.211685] env[63372]: _type = "Task" [ 816.211685] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.223193] env[63372]: DEBUG oslo_vmware.api [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1024168, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.306436] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024159, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.307635] env[63372]: DEBUG oslo_concurrency.lockutils [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.307964] env[63372]: DEBUG oslo_concurrency.lockutils [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquired lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.308258] env[63372]: DEBUG nova.network.neutron [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 816.336613] env[63372]: DEBUG nova.scheduler.client.report [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.512842] env[63372]: DEBUG oslo_vmware.api [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1024167, 'name': ReconfigVM_Task, 'duration_secs': 0.268642} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.512842] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227276', 'volume_id': '3226fd91-2732-41ce-bb8e-f909a830c8d4', 'name': 'volume-3226fd91-2732-41ce-bb8e-f909a830c8d4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '69a107d0-80c7-42e5-b514-b4273e1a3359', 'attached_at': '', 'detached_at': '', 'volume_id': '3226fd91-2732-41ce-bb8e-f909a830c8d4', 'serial': '3226fd91-2732-41ce-bb8e-f909a830c8d4'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 816.512842] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 816.512842] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a44625-281c-4e17-87ba-ec33a9efe4cd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.520468] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 816.521015] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aaa699c8-9248-4b74-ab9d-55002529aa20 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.539942] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024164, 'name': CreateVM_Task, 'duration_secs': 1.495163} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.540501] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 816.541230] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.541445] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.541987] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 816.542263] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6ed615b-17f7-44a7-9e2e-ed869ded205b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.549376] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 816.549376] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529f7728-36fa-51a4-5f66-683bed8d45b2" [ 816.549376] env[63372]: _type = "Task" [ 816.549376] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.559359] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529f7728-36fa-51a4-5f66-683bed8d45b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.605776] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 816.606042] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 816.606219] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Deleting the datastore file [datastore2] 69a107d0-80c7-42e5-b514-b4273e1a3359 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.606512] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7604dc13-a4f0-4f1d-ae56-99cbd629f9fe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.615752] env[63372]: DEBUG oslo_vmware.api [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Waiting for the task: (returnval){ [ 816.615752] env[63372]: value = "task-1024170" [ 816.615752] env[63372]: _type = "Task" [ 816.615752] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.632158] env[63372]: DEBUG oslo_vmware.api [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1024170, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.681722] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52bd7cde-ada3-55af-c9a0-eb02b57ea4b8, 'name': SearchDatastore_Task, 'duration_secs': 0.092564} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.685050] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.685379] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 357505d0-f306-4e11-8a62-e03cfab2b7c5/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk. {{(pid=63372) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 816.685976] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3075c755-b44d-4205-972d-530df510cbec {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.701998] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 816.701998] env[63372]: value = "task-1024171" [ 816.701998] env[63372]: _type = "Task" [ 816.701998] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.710056] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024161, 'name': CreateSnapshot_Task, 'duration_secs': 1.780689} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.710853] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 816.714684] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a315d4-6c90-422e-98ea-bcf0632485ff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.725269] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024171, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.731184] env[63372]: DEBUG oslo_vmware.api [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1024168, 'name': PowerOffVM_Task, 'duration_secs': 0.353065} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.735123] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 816.735365] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 816.738321] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aa0e62b2-c3ec-47d3-a314-51072d406783 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.806854] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024159, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.650747} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.807235] env[63372]: INFO nova.virt.vmwareapi.ds_util [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_81e4dbb4-984d-4a47-854d-c80457bfc995/OSTACK_IMG_81e4dbb4-984d-4a47-854d-c80457bfc995.vmdk to [datastore2] devstack-image-cache_base/3c9be794-7dcb-43df-a9b2-d9289ac8889f/3c9be794-7dcb-43df-a9b2-d9289ac8889f.vmdk. [ 816.807534] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Cleaning up location [datastore2] OSTACK_IMG_81e4dbb4-984d-4a47-854d-c80457bfc995 {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 816.807759] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_81e4dbb4-984d-4a47-854d-c80457bfc995 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.808068] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e89f3697-98ae-4a0b-b3cd-323289e113dd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.813580] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 816.813830] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 816.814040] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Deleting the datastore file [datastore1] d50919ac-3a0b-46ac-a837-ca3e6ad05173 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.814786] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d889ca5-c489-4134-8bb4-c38cf017ea5a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.819810] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 816.819810] env[63372]: value = "task-1024173" [ 816.819810] env[63372]: _type = "Task" [ 816.819810] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.824937] env[63372]: DEBUG oslo_vmware.api [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Waiting for the task: (returnval){ [ 816.824937] env[63372]: value = "task-1024174" [ 816.824937] env[63372]: _type = "Task" [ 816.824937] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.832430] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024173, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.838903] env[63372]: DEBUG oslo_vmware.api [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1024174, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.846376] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.251s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.848882] env[63372]: DEBUG oslo_concurrency.lockutils [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.632s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.849208] env[63372]: DEBUG nova.objects.instance [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lazy-loading 'resources' on Instance uuid cbd55518-a3b2-4636-ba43-c279fdd0bf8d {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 816.857461] env[63372]: DEBUG nova.network.neutron [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Updated VIF entry in instance network info cache for port c8f78ebe-604d-4414-a16d-ae25256f2730. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 816.857857] env[63372]: DEBUG nova.network.neutron [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Updating instance_info_cache with network_info: [{"id": "c8f78ebe-604d-4414-a16d-ae25256f2730", "address": "fa:16:3e:25:30:2b", "network": {"id": "73cbc616-354c-40de-971d-91fb5f0cbbac", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-817872371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2319173a3c445ca93ae2510141be32c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d", "external-id": "nsx-vlan-transportzone-591", "segmentation_id": 591, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8f78ebe-60", "ovs_interfaceid": "c8f78ebe-604d-4414-a16d-ae25256f2730", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.883795] env[63372]: INFO nova.scheduler.client.report [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Deleted allocations for instance bad91d77-c7ee-4572-b1ed-068b2a55233c [ 817.061747] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529f7728-36fa-51a4-5f66-683bed8d45b2, 'name': SearchDatastore_Task, 'duration_secs': 0.011366} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.062841] env[63372]: DEBUG nova.network.neutron [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance_info_cache with network_info: [{"id": "874f14b1-6cc3-4701-8a08-35940fc9bd9b", "address": "fa:16:3e:57:46:8e", "network": {"id": "2c9e4152-311b-4f28-9bf0-323ad33d6136", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.24", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "93ad9fbc5a8646d6b4cf17fa4ff19fe2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6ca3b2e-69a5-4cea-96a7-eaad5ec5fd9b", "external-id": "nsx-vlan-transportzone-989", "segmentation_id": 989, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap874f14b1-6c", "ovs_interfaceid": "874f14b1-6cc3-4701-8a08-35940fc9bd9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.064181] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.064331] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.064620] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.064773] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.064957] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.065544] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a4e4f9b-19c1-4d88-8ff1-7542b46ea5a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.081294] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.081580] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 817.082787] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-908ba73c-cac6-42e2-81ea-451ad16d8c7e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.091744] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 817.091744] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52753b04-4cab-43f3-30fa-82c0618febfd" [ 817.091744] env[63372]: _type = "Task" [ 817.091744] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.101962] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52753b04-4cab-43f3-30fa-82c0618febfd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.130992] env[63372]: DEBUG oslo_vmware.api [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Task: {'id': task-1024170, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09768} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.130992] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 817.130992] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 817.130992] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 817.130992] env[63372]: INFO nova.compute.manager [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Took 2.37 seconds to destroy the instance on the hypervisor. [ 817.130992] env[63372]: DEBUG oslo.service.loopingcall [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.130992] env[63372]: DEBUG nova.compute.manager [-] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 817.130992] env[63372]: DEBUG nova.network.neutron [-] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 817.217174] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024171, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.248142] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 817.248584] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ad846f66-58e3-4e88-b5ad-9e76a403e778 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.259191] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 817.259191] env[63372]: value = "task-1024175" [ 817.259191] env[63372]: _type = "Task" [ 817.259191] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.271316] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024175, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.334717] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024173, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.046772} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.335444] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 817.335641] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3c9be794-7dcb-43df-a9b2-d9289ac8889f/3c9be794-7dcb-43df-a9b2-d9289ac8889f.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.336031] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3c9be794-7dcb-43df-a9b2-d9289ac8889f/3c9be794-7dcb-43df-a9b2-d9289ac8889f.vmdk to [datastore2] 368101ee-e4fd-4fe4-b7c8-75c6b210189f/368101ee-e4fd-4fe4-b7c8-75c6b210189f.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 817.336309] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-139bc932-b7e4-40a5-aa89-90b2da3afc6b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.342955] env[63372]: DEBUG oslo_vmware.api [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1024174, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.349506] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 817.349506] env[63372]: value = "task-1024176" [ 817.349506] env[63372]: _type = "Task" [ 817.349506] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.361872] env[63372]: DEBUG oslo_concurrency.lockutils [req-a107544b-2145-4a96-b1ac-01f8dcbc65ae req-be50c44d-cf3d-4255-bf73-79f933c19deb service nova] Releasing lock "refresh_cache-d50919ac-3a0b-46ac-a837-ca3e6ad05173" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.367899] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024176, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.395739] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e90f6ea4-1e81-4cbf-9a04-a95207a0bb7a tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "bad91d77-c7ee-4572-b1ed-068b2a55233c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.303s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.567826] env[63372]: DEBUG oslo_concurrency.lockutils [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Releasing lock "refresh_cache-8283b736-ad02-4082-97b7-561bd5c5da93" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.567826] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 817.573222] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb2e00de-ad13-49ca-b8a9-6d6944cc836b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.583264] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 817.583264] env[63372]: value = "task-1024177" [ 817.583264] env[63372]: _type = "Task" [ 817.583264] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.603430] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024177, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.611917] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52753b04-4cab-43f3-30fa-82c0618febfd, 'name': SearchDatastore_Task, 'duration_secs': 0.067011} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.612916] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd7ed230-33c2-4ff2-895e-d992316a3d46 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.627025] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 817.627025] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fc1fd0-b7b4-0ce8-2b9a-ba8f5c1a7637" [ 817.627025] env[63372]: _type = "Task" [ 817.627025] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.636672] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fc1fd0-b7b4-0ce8-2b9a-ba8f5c1a7637, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.717340] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024171, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.762312} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.721217] env[63372]: INFO nova.virt.vmwareapi.ds_util [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 357505d0-f306-4e11-8a62-e03cfab2b7c5/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk. [ 817.723575] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f228c41-4889-49dd-b645-f10128155255 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.756311] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 357505d0-f306-4e11-8a62-e03cfab2b7c5/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.759750] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46017b54-f87b-4c18-bf07-fa8ab49e4b02 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.786415] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024175, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.788681] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 817.788681] env[63372]: value = "task-1024178" [ 817.788681] env[63372]: _type = "Task" [ 817.788681] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.805259] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024178, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.837704] env[63372]: DEBUG oslo_vmware.api [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Task: {'id': task-1024174, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.62085} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.837704] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 817.838118] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 817.838262] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 817.838552] env[63372]: INFO nova.compute.manager [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Took 1.65 seconds to destroy the instance on the hypervisor. [ 817.838970] env[63372]: DEBUG oslo.service.loopingcall [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.839288] env[63372]: DEBUG nova.compute.manager [-] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 817.839431] env[63372]: DEBUG nova.network.neutron [-] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 817.863351] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024176, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.923369] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77989400-9dd1-417b-a035-28222d9e12bf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.934963] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e55dc817-f4d2-4ab3-922c-e11b14708777 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.971100] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f131c764-0f54-4abb-b1de-1a10bb192966 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.981117] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6aac6e8-c88c-40e1-ae3c-02d04a9adc42 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.996759] env[63372]: DEBUG nova.compute.provider_tree [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.100843] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024177, 'name': PowerOffVM_Task, 'duration_secs': 0.262323} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.100843] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 818.102683] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:36:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='51633402-95cc-485c-8808-65539d485326',id=27,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-443337402',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.102683] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.102683] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.102683] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.102683] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.102683] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.102683] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.103114] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.103114] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.103207] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.103390] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.109677] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c53e5574-4091-42dc-85b1-108556d88e4d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.133420] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 818.133420] env[63372]: value = "task-1024179" [ 818.133420] env[63372]: _type = "Task" [ 818.133420] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.143411] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fc1fd0-b7b4-0ce8-2b9a-ba8f5c1a7637, 'name': SearchDatastore_Task, 'duration_secs': 0.016466} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.144171] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.144447] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 89301344-84a9-4d13-aae7-99943d0a478e/89301344-84a9-4d13-aae7-99943d0a478e.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 818.145053] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5819e74-388b-40bc-859a-020c254ffc00 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.152904] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024179, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.155772] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 818.155772] env[63372]: value = "task-1024180" [ 818.155772] env[63372]: _type = "Task" [ 818.155772] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.167186] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024180, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.193456] env[63372]: DEBUG nova.network.neutron [-] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.226916] env[63372]: DEBUG nova.compute.manager [req-13c53455-157d-4d5c-adcb-a6a699bcb345 req-f92412b8-9918-4254-a77a-3874ff92034f service nova] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Received event network-vif-deleted-29995495-a43f-4be1-b907-c88d9fe78dcc {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.289859] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024175, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.305039] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024178, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.368021] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024176, 'name': CopyVirtualDisk_Task} progress is 32%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.500081] env[63372]: DEBUG nova.scheduler.client.report [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.646891] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024179, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.676378] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024180, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.698233] env[63372]: INFO nova.compute.manager [-] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Took 1.57 seconds to deallocate network for instance. [ 818.789018] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024175, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.805151] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024178, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.862636] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024176, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.934312] env[63372]: DEBUG nova.network.neutron [-] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.006698] env[63372]: DEBUG oslo_concurrency.lockutils [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.158s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.009897] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 25.844s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.041280] env[63372]: INFO nova.scheduler.client.report [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Deleted allocations for instance cbd55518-a3b2-4636-ba43-c279fdd0bf8d [ 819.146238] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024179, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.173847] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024180, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.266471] env[63372]: INFO nova.compute.manager [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Took 0.57 seconds to detach 1 volumes for instance. [ 819.270853] env[63372]: DEBUG nova.compute.manager [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Deleting volume: 3226fd91-2732-41ce-bb8e-f909a830c8d4 {{(pid=63372) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 819.288046] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024175, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.309075] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024178, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.370156] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024176, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.443909] env[63372]: INFO nova.compute.manager [-] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Took 1.60 seconds to deallocate network for instance. [ 819.562465] env[63372]: DEBUG oslo_concurrency.lockutils [None req-103f6c86-80ab-4459-b72a-f1e4be34627c tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "cbd55518-a3b2-4636-ba43-c279fdd0bf8d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.902s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.653061] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024179, 'name': ReconfigVM_Task, 'duration_secs': 1.027874} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.653061] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f471218-ec76-4283-8484-fcc846b90951 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.676219] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:36:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='51633402-95cc-485c-8808-65539d485326',id=27,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-443337402',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 819.676493] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 819.676914] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.676914] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 819.678125] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.678307] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 819.678610] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 819.679035] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 819.679035] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 819.679229] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 819.679381] env[63372]: DEBUG nova.virt.hardware [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.683965] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54a76a4b-406c-41cc-be1d-4d290dd03077 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.691522] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 819.691522] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525777a6-9afd-2049-6968-e6f6a938d29b" [ 819.691522] env[63372]: _type = "Task" [ 819.691522] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.696019] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024180, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.709867] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525777a6-9afd-2049-6968-e6f6a938d29b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.792441] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024175, 'name': CloneVM_Task} progress is 95%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.805493] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024178, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.829845] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.866129] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024176, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.955563] env[63372]: DEBUG oslo_concurrency.lockutils [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.060339] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance cf673ac1-2c7d-468b-83ec-c723d5182457 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.060564] env[63372]: WARNING nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance bbba9f28-045d-41ab-8539-5b2968fe3d54 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.060735] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance ac90a156-be00-4f62-a76e-e08914531167 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.060998] env[63372]: WARNING nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 5819c38e-2cf9-4d16-b28a-5f23d35c3d44 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.061163] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance c4718797-aa86-4ec0-94d3-6480bd6aa898 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.061370] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.061545] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance d50919ac-3a0b-46ac-a837-ca3e6ad05173 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.061732] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance b182294d-2de8-4189-af7f-3e2d2c604a8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.061934] env[63372]: WARNING nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance f0c60559-c072-4b61-afe8-03d6c131b307 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.062104] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 93a5d948-0629-4f53-a681-858d519acfa7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.062284] env[63372]: WARNING nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance f43555ef-b517-4b7d-9d2e-4787d40e201b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.062457] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance cd0c01ac-602b-44a3-8099-84b8a50b2449 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.062585] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance ae6f3f42-7213-4ab1-b74c-1a557df6748b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.062732] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 69a107d0-80c7-42e5-b514-b4273e1a3359 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.062876] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 3f66eda6-f5e9-4527-9711-849a01702580 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.063038] env[63372]: WARNING nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance f6675697-1529-46be-b28a-398ff3060d18 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.063167] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 8283b736-ad02-4082-97b7-561bd5c5da93 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.063281] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 99f901a6-9bb3-4403-af0c-c8900f655cb3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.063390] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance de1b38a4-c7f2-420c-a050-7311976e4ca8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.063525] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 7c9d551e-d210-4943-971f-b2829751fcfc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.063632] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance c01a5d24-eb46-4a69-993e-753880ce8e85 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.063737] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 357505d0-f306-4e11-8a62-e03cfab2b7c5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.063866] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 368101ee-e4fd-4fe4-b7c8-75c6b210189f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.063973] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 89301344-84a9-4d13-aae7-99943d0a478e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 820.187326] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024180, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.208091] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525777a6-9afd-2049-6968-e6f6a938d29b, 'name': SearchDatastore_Task, 'duration_secs': 0.044706} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.215158] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Reconfiguring VM instance instance-0000002f to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 820.215717] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6eb1c07-70f2-4343-a28c-99293b172bef {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.254770] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 820.254770] env[63372]: value = "task-1024182" [ 820.254770] env[63372]: _type = "Task" [ 820.254770] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.265956] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024182, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.290446] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024175, 'name': CloneVM_Task, 'duration_secs': 2.696428} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.290723] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Created linked-clone VM from snapshot [ 820.292811] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0642aa-d211-4ea2-b8d9-cb8aa64ca832 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.302471] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Uploading image 77e136a4-4a74-4ff7-8ff7-0eb4fa83bb02 {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 820.315260] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024178, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.333302] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 820.333302] env[63372]: value = "vm-227370" [ 820.333302] env[63372]: _type = "VirtualMachine" [ 820.333302] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 820.333642] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c0789f83-6bb9-49d4-90f6-3e5c5d88e2a2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.343333] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lease: (returnval){ [ 820.343333] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e1cd90-bed7-d581-9084-013ee84e8895" [ 820.343333] env[63372]: _type = "HttpNfcLease" [ 820.343333] env[63372]: } obtained for exporting VM: (result){ [ 820.343333] env[63372]: value = "vm-227370" [ 820.343333] env[63372]: _type = "VirtualMachine" [ 820.343333] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 820.343669] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the lease: (returnval){ [ 820.343669] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e1cd90-bed7-d581-9084-013ee84e8895" [ 820.343669] env[63372]: _type = "HttpNfcLease" [ 820.343669] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 820.351815] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 820.351815] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e1cd90-bed7-d581-9084-013ee84e8895" [ 820.351815] env[63372]: _type = "HttpNfcLease" [ 820.351815] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 820.364219] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024176, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.715919} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.364219] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3c9be794-7dcb-43df-a9b2-d9289ac8889f/3c9be794-7dcb-43df-a9b2-d9289ac8889f.vmdk to [datastore2] 368101ee-e4fd-4fe4-b7c8-75c6b210189f/368101ee-e4fd-4fe4-b7c8-75c6b210189f.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 820.364909] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64742e9-bd84-4ddb-b5de-72f04b68e107 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.389420] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 368101ee-e4fd-4fe4-b7c8-75c6b210189f/368101ee-e4fd-4fe4-b7c8-75c6b210189f.vmdk or device None with type streamOptimized {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.389765] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-117ec535-7a52-4d9f-b39e-a5763a224c12 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.413016] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 820.413016] env[63372]: value = "task-1024184" [ 820.413016] env[63372]: _type = "Task" [ 820.413016] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.423344] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024184, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.473647] env[63372]: DEBUG oslo_concurrency.lockutils [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "7c9d551e-d210-4943-971f-b2829751fcfc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.474211] env[63372]: DEBUG oslo_concurrency.lockutils [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "7c9d551e-d210-4943-971f-b2829751fcfc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.474211] env[63372]: DEBUG oslo_concurrency.lockutils [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "7c9d551e-d210-4943-971f-b2829751fcfc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.474383] env[63372]: DEBUG oslo_concurrency.lockutils [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "7c9d551e-d210-4943-971f-b2829751fcfc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.474971] env[63372]: DEBUG oslo_concurrency.lockutils [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "7c9d551e-d210-4943-971f-b2829751fcfc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.477064] env[63372]: INFO nova.compute.manager [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Terminating instance [ 820.484515] env[63372]: DEBUG nova.compute.manager [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 820.484515] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 820.485571] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d253402-4fca-466e-a63b-dcb16daa773d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.492131] env[63372]: DEBUG nova.compute.manager [req-825e9fd7-3320-4c0e-a4bf-8a0c84461bc1 req-696c6f9c-2dcb-46d8-a95f-1928f8b7848e service nova] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Received event network-vif-deleted-c8f78ebe-604d-4414-a16d-ae25256f2730 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.498774] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 820.499063] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09d7b65e-8141-44ee-ba94-67a0f210ef3b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.509269] env[63372]: DEBUG oslo_vmware.api [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 820.509269] env[63372]: value = "task-1024185" [ 820.509269] env[63372]: _type = "Task" [ 820.509269] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.518624] env[63372]: DEBUG oslo_vmware.api [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024185, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.567105] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 4028d7c0-f398-4e43-9cff-5d89a14c4efa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 820.694253] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024180, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.135144} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.694253] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 89301344-84a9-4d13-aae7-99943d0a478e/89301344-84a9-4d13-aae7-99943d0a478e.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 820.694253] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.694253] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4ba0e52e-ff2c-4095-9053-9c054c25ea4f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.701183] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 820.701183] env[63372]: value = "task-1024186" [ 820.701183] env[63372]: _type = "Task" [ 820.701183] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.711500] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024186, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.767786] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024182, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.815638] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024178, 'name': ReconfigVM_Task, 'duration_secs': 2.551042} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.815903] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 357505d0-f306-4e11-8a62-e03cfab2b7c5/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.816811] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4988e92-1590-46d0-b93e-b9f6a75fd6df {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.846777] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4494bb6-21bb-45ab-b68a-cd7068fcfacb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.865204] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 820.865204] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e1cd90-bed7-d581-9084-013ee84e8895" [ 820.865204] env[63372]: _type = "HttpNfcLease" [ 820.865204] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 820.865440] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 820.865440] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e1cd90-bed7-d581-9084-013ee84e8895" [ 820.865440] env[63372]: _type = "HttpNfcLease" [ 820.865440] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 820.866228] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d4fd00-30a1-4b4d-8905-79009526d933 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.869976] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 820.869976] env[63372]: value = "task-1024187" [ 820.869976] env[63372]: _type = "Task" [ 820.869976] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.876998] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d21b3e-b8d2-3614-b4cf-1ac8c9255696/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 820.877215] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d21b3e-b8d2-3614-b4cf-1ac8c9255696/disk-0.vmdk for reading. {{(pid=63372) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 820.885957] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024187, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.952244] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024184, 'name': ReconfigVM_Task, 'duration_secs': 0.481657} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.952552] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 368101ee-e4fd-4fe4-b7c8-75c6b210189f/368101ee-e4fd-4fe4-b7c8-75c6b210189f.vmdk or device None with type streamOptimized {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.953298] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-76aea768-029c-4ce9-93f1-f8253f10af37 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.961689] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 820.961689] env[63372]: value = "task-1024188" [ 820.961689] env[63372]: _type = "Task" [ 820.961689] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.971444] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024188, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.980648] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3d8ba241-3ded-4f3d-aa4e-373ca8fbd1ba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.023471] env[63372]: DEBUG oslo_vmware.api [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024185, 'name': PowerOffVM_Task, 'duration_secs': 0.251599} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.029100] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 821.029513] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 821.029744] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-96b0995f-f04d-4d75-a62f-cb37f27de723 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.070945] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance a3d5b93e-1b76-48e7-bb3e-2423e7c4119a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 821.071314] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Total usable vcpus: 48, total allocated vcpus: 19 {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 821.071515] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4224MB phys_disk=200GB used_disk=18GB total_vcpus=48 used_vcpus=19 pci_stats=[] {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 821.078308] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Acquiring lock "07dab964-a741-4723-98a3-0cfbe4eff975" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.078600] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Lock "07dab964-a741-4723-98a3-0cfbe4eff975" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.106036] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 821.106308] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 821.106493] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Deleting the datastore file [datastore1] 7c9d551e-d210-4943-971f-b2829751fcfc {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 821.106770] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ed536ed1-ec0c-4dc6-a17e-ca11768bec75 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.118592] env[63372]: DEBUG oslo_vmware.api [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 821.118592] env[63372]: value = "task-1024190" [ 821.118592] env[63372]: _type = "Task" [ 821.118592] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.131603] env[63372]: DEBUG oslo_vmware.api [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024190, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.224483] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024186, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085895} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.224921] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 821.226175] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c524326c-3a41-4c72-8182-14b7a85f688e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.255667] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 89301344-84a9-4d13-aae7-99943d0a478e/89301344-84a9-4d13-aae7-99943d0a478e.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.257013] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-696cf398-77b9-4a5f-b5af-0a2da032d38f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.291122] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024182, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.291122] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 821.291122] env[63372]: value = "task-1024191" [ 821.291122] env[63372]: _type = "Task" [ 821.291122] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.301739] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024191, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.145392] env[63372]: DEBUG nova.compute.manager [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 822.170103] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "798047e4-1a2b-458c-86c2-5567197d1861" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.170395] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "798047e4-1a2b-458c-86c2-5567197d1861" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.172345] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "1d047728-50d8-465c-b217-6cbe9bbea3e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.172584] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "1d047728-50d8-465c-b217-6cbe9bbea3e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.182997] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024187, 'name': ReconfigVM_Task, 'duration_secs': 0.775087} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.190105] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 822.190105] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3d1cee9-4609-4ac7-8111-ec7b98e47b3d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.201932] env[63372]: DEBUG oslo_vmware.api [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024190, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169007} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.202295] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024191, 'name': ReconfigVM_Task, 'duration_secs': 0.527332} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.202588] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024188, 'name': Rename_Task, 'duration_secs': 0.190778} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.202991] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024182, 'name': ReconfigVM_Task, 'duration_secs': 1.244471} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.208345] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 822.208774] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 822.208774] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 822.209758] env[63372]: INFO nova.compute.manager [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Took 1.72 seconds to destroy the instance on the hypervisor. [ 822.209758] env[63372]: DEBUG oslo.service.loopingcall [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.209758] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 89301344-84a9-4d13-aae7-99943d0a478e/89301344-84a9-4d13-aae7-99943d0a478e.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 822.210244] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 822.210539] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Reconfigured VM instance instance-0000002f to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 822.211473] env[63372]: DEBUG nova.compute.manager [-] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 822.211774] env[63372]: DEBUG nova.network.neutron [-] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 822.213438] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-21a838c0-fdb9-4196-913e-1f4465662e70 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.215194] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2cce55d2-3aa9-4d70-9ec7-2ff5de04f271 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.217908] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7501b9e6-6a15-4852-a5ce-334a38fb3181 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.222170] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 822.222170] env[63372]: value = "task-1024192" [ 822.222170] env[63372]: _type = "Task" [ 822.222170] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.248820] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] 8283b736-ad02-4082-97b7-561bd5c5da93/8283b736-ad02-4082-97b7-561bd5c5da93.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.255406] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b358785f-a2b3-48ed-af6d-2d1d9dab1ee3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.272641] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 822.272641] env[63372]: value = "task-1024194" [ 822.272641] env[63372]: _type = "Task" [ 822.272641] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.272641] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 822.272641] env[63372]: value = "task-1024193" [ 822.272641] env[63372]: _type = "Task" [ 822.272641] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.277492] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024192, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.287276] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 822.287276] env[63372]: value = "task-1024195" [ 822.287276] env[63372]: _type = "Task" [ 822.287276] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.299893] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024194, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.300323] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024193, 'name': Rename_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.309498] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.411973] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8116e59-a7b3-44a0-b718-8b38d6e0a04f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.426512] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9049ea01-6967-4c9e-b547-107ef4c6bca6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.468726] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f75457b6-cb9d-45f6-8fb1-a689ff14f212 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.478082] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b4b767-c021-4eca-a559-91944498e787 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.494597] env[63372]: DEBUG nova.compute.provider_tree [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.654203] env[63372]: DEBUG nova.compute.manager [req-04db2e54-144d-4216-9f2a-3a2e228c9258 req-2c92b92e-768e-4719-890c-b212520f6eaf service nova] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Received event network-vif-deleted-a0cd67cc-8bd5-4d24-8620-042e1fecd124 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.654434] env[63372]: INFO nova.compute.manager [req-04db2e54-144d-4216-9f2a-3a2e228c9258 req-2c92b92e-768e-4719-890c-b212520f6eaf service nova] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Neutron deleted interface a0cd67cc-8bd5-4d24-8620-042e1fecd124; detaching it from the instance and deleting it from the info cache [ 822.656853] env[63372]: DEBUG nova.network.neutron [req-04db2e54-144d-4216-9f2a-3a2e228c9258 req-2c92b92e-768e-4719-890c-b212520f6eaf service nova] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.673081] env[63372]: DEBUG nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 822.678062] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.678893] env[63372]: DEBUG nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 822.737780] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024192, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.797925] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024193, 'name': Rename_Task, 'duration_secs': 0.234977} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.798483] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024194, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.799281] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 822.799621] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d22f850-922d-4872-9534-e9a15fe434a1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.804373] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024195, 'name': ReconfigVM_Task, 'duration_secs': 0.496555} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.805095] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Reconfigured VM instance instance-0000002f to attach disk [datastore2] 8283b736-ad02-4082-97b7-561bd5c5da93/8283b736-ad02-4082-97b7-561bd5c5da93.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 822.806249] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dc3cc68-ed49-454b-8dd0-36e896b11a8c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.810903] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 822.810903] env[63372]: value = "task-1024196" [ 822.810903] env[63372]: _type = "Task" [ 822.810903] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.830582] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9e343c-3735-4a01-b377-2780e2b09658 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.855815] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6acaaed9-e5f6-4c3f-a9f5-d7c9397816cf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.859066] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024196, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.878737] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8bfeb8-debc-4ebc-bd98-16bf5ca299c5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.887527] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 822.887987] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9728f6f0-6e1a-4e56-b076-d00d206e8fc6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.896086] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 822.896086] env[63372]: value = "task-1024197" [ 822.896086] env[63372]: _type = "Task" [ 822.896086] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.906382] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024197, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.998298] env[63372]: DEBUG nova.scheduler.client.report [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.118628] env[63372]: DEBUG nova.network.neutron [-] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.161369] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-486e4eb1-6e6b-478b-9aca-d11ecbad6a1a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.173060] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87986d05-826b-4639-8f69-d02bf6237c49 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.205715] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.207099] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.223742] env[63372]: DEBUG nova.compute.manager [req-04db2e54-144d-4216-9f2a-3a2e228c9258 req-2c92b92e-768e-4719-890c-b212520f6eaf service nova] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Detach interface failed, port_id=a0cd67cc-8bd5-4d24-8620-042e1fecd124, reason: Instance 7c9d551e-d210-4943-971f-b2829751fcfc could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 823.236513] env[63372]: DEBUG oslo_vmware.api [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024192, 'name': PowerOnVM_Task, 'duration_secs': 0.830009} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.236625] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 823.239935] env[63372]: DEBUG nova.compute.manager [None req-a96a35b6-a085-45c0-977f-15a5f5b6d9a1 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.240854] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd646f6-2ae0-447a-b90d-fbbcc52cf257 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.291300] env[63372]: DEBUG oslo_vmware.api [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024194, 'name': PowerOnVM_Task, 'duration_secs': 0.680564} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.291657] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 823.291908] env[63372]: INFO nova.compute.manager [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Took 19.30 seconds to spawn the instance on the hypervisor. [ 823.292132] env[63372]: DEBUG nova.compute.manager [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.292939] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811575cd-b647-4b8f-b074-8486e31bf732 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.323514] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024196, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.409877] env[63372]: DEBUG oslo_vmware.api [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024197, 'name': PowerOnVM_Task, 'duration_secs': 0.478269} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.410327] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 823.506493] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 823.506825] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.497s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.507219] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.427s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.510605] env[63372]: INFO nova.compute.claims [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.519231] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 823.519609] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Cleaning up deleted instances {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 823.621051] env[63372]: INFO nova.compute.manager [-] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Took 1.41 seconds to deallocate network for instance. [ 823.812148] env[63372]: INFO nova.compute.manager [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Took 44.04 seconds to build instance. [ 823.823616] env[63372]: DEBUG oslo_vmware.api [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024196, 'name': PowerOnVM_Task, 'duration_secs': 0.785591} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.824632] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 823.824915] env[63372]: INFO nova.compute.manager [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Took 12.58 seconds to spawn the instance on the hypervisor. [ 823.825202] env[63372]: DEBUG nova.compute.manager [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.826367] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af9be37-43ba-4b4d-abc4-e83ef03aea2e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.032801] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] There are 19 instances to clean {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 824.033271] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: cbd55518-a3b2-4636-ba43-c279fdd0bf8d] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 824.128713] env[63372]: DEBUG oslo_concurrency.lockutils [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.266273] env[63372]: INFO nova.compute.manager [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Unrescuing [ 824.266691] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "refresh_cache-357505d0-f306-4e11-8a62-e03cfab2b7c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.266691] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquired lock "refresh_cache-357505d0-f306-4e11-8a62-e03cfab2b7c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.266845] env[63372]: DEBUG nova.network.neutron [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 824.317979] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6f88f2d9-53c0-45d5-a357-dae7cc056512 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "368101ee-e4fd-4fe4-b7c8-75c6b210189f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.425s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.350642] env[63372]: INFO nova.compute.manager [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Took 39.52 seconds to build instance. [ 824.429078] env[63372]: INFO nova.compute.manager [None req-151087bc-658c-42b7-bd69-355cf1a32fd0 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance to original state: 'active' [ 824.540756] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: bad91d77-c7ee-4572-b1ed-068b2a55233c] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 824.853712] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4d5baab7-c073-4e85-a7c8-63f61e1ef352 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "89301344-84a9-4d13-aae7-99943d0a478e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.842s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.947983] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcd7d13-c7b0-4fa1-b0a9-3505b191cae0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.956650] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f023bf-46b7-441e-9c62-6347bca58635 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.993206] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b053a9-139c-4d42-892e-b6f9c01ce1d9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.005159] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c7ca99-354d-4f7c-b001-6981a720c26d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.021559] env[63372]: DEBUG nova.compute.provider_tree [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.037210] env[63372]: DEBUG nova.network.neutron [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Updating instance_info_cache with network_info: [{"id": "bf61306f-1264-4480-b485-299826369f7d", "address": "fa:16:3e:13:43:f9", "network": {"id": "2f241a45-cc18-46f6-b0a3-1aad8d2713fd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1953733697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a14a1aa8a0574b65a312754ae1091f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a88b707-352e-4be7-b1d6-ad6074b40ed9", "external-id": "nsx-vlan-transportzone-789", "segmentation_id": 789, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf61306f-12", "ovs_interfaceid": "bf61306f-1264-4480-b485-299826369f7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.044158] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 4c9eb955-3bed-4b26-866f-b2a876b835dd] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 825.276618] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "368101ee-e4fd-4fe4-b7c8-75c6b210189f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.277598] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "368101ee-e4fd-4fe4-b7c8-75c6b210189f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.277598] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "368101ee-e4fd-4fe4-b7c8-75c6b210189f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.277761] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "368101ee-e4fd-4fe4-b7c8-75c6b210189f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.278061] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "368101ee-e4fd-4fe4-b7c8-75c6b210189f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.280808] env[63372]: INFO nova.compute.manager [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Terminating instance [ 825.282971] env[63372]: DEBUG nova.compute.manager [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 825.283314] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 825.284388] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0596d2e-710e-4ebb-aec9-1770e991601d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.296684] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 825.297229] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e3689a9-5298-44cb-ae4f-2ff7a59c6567 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.306339] env[63372]: DEBUG oslo_vmware.api [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 825.306339] env[63372]: value = "task-1024198" [ 825.306339] env[63372]: _type = "Task" [ 825.306339] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.318419] env[63372]: DEBUG oslo_vmware.api [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024198, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.526255] env[63372]: DEBUG nova.scheduler.client.report [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.540082] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Releasing lock "refresh_cache-357505d0-f306-4e11-8a62-e03cfab2b7c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.541256] env[63372]: DEBUG nova.objects.instance [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lazy-loading 'flavor' on Instance uuid 357505d0-f306-4e11-8a62-e03cfab2b7c5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 825.548796] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: febeca1c-3370-4c74-8bd9-efad313df1e2] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 825.735179] env[63372]: DEBUG nova.compute.manager [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 825.736687] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66dd5a57-22f1-49dc-b7a2-aced21a1ebf0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.772520] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "8283b736-ad02-4082-97b7-561bd5c5da93" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.772826] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "8283b736-ad02-4082-97b7-561bd5c5da93" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.773072] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "8283b736-ad02-4082-97b7-561bd5c5da93-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.773281] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "8283b736-ad02-4082-97b7-561bd5c5da93-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.773886] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "8283b736-ad02-4082-97b7-561bd5c5da93-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.776963] env[63372]: INFO nova.compute.manager [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Terminating instance [ 825.779346] env[63372]: DEBUG nova.compute.manager [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 825.779569] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 825.780478] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-185d761d-5643-4b78-8b34-315ed7076f92 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.791019] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 825.791294] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41e08657-59a2-4dbb-b3ff-c28e61110849 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.799820] env[63372]: DEBUG oslo_vmware.api [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 825.799820] env[63372]: value = "task-1024199" [ 825.799820] env[63372]: _type = "Task" [ 825.799820] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.810102] env[63372]: DEBUG oslo_vmware.api [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024199, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.820119] env[63372]: DEBUG oslo_vmware.api [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024198, 'name': PowerOffVM_Task, 'duration_secs': 0.239251} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.820423] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 825.820583] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 825.820835] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0fb4c928-a323-4656-9403-539de1ca91e2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.032041] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.525s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.032713] env[63372]: DEBUG nova.compute.manager [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 826.036128] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.673s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.037706] env[63372]: INFO nova.compute.claims [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.052651] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624c9583-017b-455a-a8fd-03a1d0b55aa7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.056333] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 757df632-4af2-4eb3-bc6d-8c0812ba850d] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 826.077302] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 826.077730] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6b4131d9-7f46-4ab9-82fd-61cb07010f30 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.087393] env[63372]: DEBUG oslo_vmware.api [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 826.087393] env[63372]: value = "task-1024201" [ 826.087393] env[63372]: _type = "Task" [ 826.087393] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.099184] env[63372]: DEBUG oslo_vmware.api [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024201, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.250248] env[63372]: INFO nova.compute.manager [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] instance snapshotting [ 826.254485] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2607a6e-6474-4219-a7f5-c70f10d43115 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.279141] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328dc786-e935-47d7-906a-92a29aaa3ed6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.314758] env[63372]: DEBUG oslo_vmware.api [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024199, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.497443] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 826.497760] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 826.497927] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Deleting the datastore file [datastore2] 368101ee-e4fd-4fe4-b7c8-75c6b210189f {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 826.498266] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-742076e8-6dd4-4d17-a042-d9bb5a39d228 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.508020] env[63372]: DEBUG oslo_vmware.api [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 826.508020] env[63372]: value = "task-1024202" [ 826.508020] env[63372]: _type = "Task" [ 826.508020] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.518167] env[63372]: DEBUG oslo_vmware.api [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024202, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.543434] env[63372]: DEBUG nova.compute.utils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.548659] env[63372]: DEBUG nova.compute.manager [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.548659] env[63372]: DEBUG nova.network.neutron [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 826.559156] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 7cf6ac9b-4703-4143-a22c-abc5528ce5ef] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 826.598906] env[63372]: DEBUG oslo_vmware.api [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024201, 'name': PowerOffVM_Task, 'duration_secs': 0.399187} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.599285] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 826.605158] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Reconfiguring VM instance instance-00000042 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 826.606854] env[63372]: DEBUG nova.policy [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb233d4b6fe54fefae3046d62d167e34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d0c0f315749429dbae22a72ae82b1b2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.608782] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46cacdd9-23cf-44b0-b819-6c8034171ab0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.629518] env[63372]: DEBUG oslo_vmware.api [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 826.629518] env[63372]: value = "task-1024203" [ 826.629518] env[63372]: _type = "Task" [ 826.629518] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.639554] env[63372]: DEBUG oslo_vmware.api [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024203, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.792533] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 826.792849] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-dc748bab-09dd-41b1-babd-fdd2c4504c6a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.802341] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 826.802341] env[63372]: value = "task-1024204" [ 826.802341] env[63372]: _type = "Task" [ 826.802341] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.814771] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024204, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.817784] env[63372]: DEBUG oslo_vmware.api [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024199, 'name': PowerOffVM_Task, 'duration_secs': 0.681292} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.818109] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 826.818382] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 826.818660] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-98db57ed-113e-4365-83f2-cf4d5c4557bc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.884797] env[63372]: DEBUG nova.network.neutron [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Successfully created port: 95bb06d1-3333-43b4-a667-f99370acc1d7 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.914354] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 826.914523] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 826.914756] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Deleting the datastore file [datastore2] 8283b736-ad02-4082-97b7-561bd5c5da93 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 826.915078] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14ea0187-2587-4d06-9625-9742e1fad971 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.923809] env[63372]: DEBUG oslo_vmware.api [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for the task: (returnval){ [ 826.923809] env[63372]: value = "task-1024206" [ 826.923809] env[63372]: _type = "Task" [ 826.923809] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.934660] env[63372]: DEBUG oslo_vmware.api [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024206, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.023998] env[63372]: DEBUG oslo_vmware.api [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024202, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.4411} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.026021] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 827.026021] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 827.026021] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 827.026021] env[63372]: INFO nova.compute.manager [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Took 1.74 seconds to destroy the instance on the hypervisor. [ 827.026021] env[63372]: DEBUG oslo.service.loopingcall [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.026021] env[63372]: DEBUG nova.compute.manager [-] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 827.026021] env[63372]: DEBUG nova.network.neutron [-] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 827.059206] env[63372]: DEBUG nova.compute.manager [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.069577] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 7a75a51d-b804-453b-ba7b-fda6c4931802] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 827.145048] env[63372]: DEBUG oslo_vmware.api [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024203, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.320148] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024204, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.438723] env[63372]: DEBUG oslo_vmware.api [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Task: {'id': task-1024206, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.330204} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.438998] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 827.439206] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 827.439387] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 827.439760] env[63372]: INFO nova.compute.manager [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Took 1.66 seconds to destroy the instance on the hypervisor. [ 827.439843] env[63372]: DEBUG oslo.service.loopingcall [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.440030] env[63372]: DEBUG nova.compute.manager [-] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 827.440129] env[63372]: DEBUG nova.network.neutron [-] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 827.513187] env[63372]: DEBUG nova.compute.manager [req-d44ccfb2-a816-4a85-9734-f245cb0a44ee req-626b4ad2-8065-47d6-aaad-b2f6805fa993 service nova] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Received event network-vif-deleted-c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.513508] env[63372]: INFO nova.compute.manager [req-d44ccfb2-a816-4a85-9734-f245cb0a44ee req-626b4ad2-8065-47d6-aaad-b2f6805fa993 service nova] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Neutron deleted interface c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa; detaching it from the instance and deleting it from the info cache [ 827.513763] env[63372]: DEBUG nova.network.neutron [req-d44ccfb2-a816-4a85-9734-f245cb0a44ee req-626b4ad2-8065-47d6-aaad-b2f6805fa993 service nova] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.517340] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01899fb2-41bf-43f5-bca2-21a98d9d2d47 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.528479] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b134071c-ce6f-4b8d-beac-b0e741c76220 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.562207] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d0460a5-d93d-4519-bf04-c6004383cbd0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.575158] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 63b9095c-fb49-4dc1-a6dc-96529aeaab81] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 827.579982] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a84bf73-eeed-4cb3-9f0d-3d013510db05 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.599732] env[63372]: DEBUG nova.compute.provider_tree [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.642036] env[63372]: DEBUG oslo_vmware.api [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024203, 'name': ReconfigVM_Task, 'duration_secs': 0.52649} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.642036] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Reconfigured VM instance instance-00000042 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 827.642036] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 827.642968] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7f7c26f-8744-43f0-b503-a8e8c823885e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.651738] env[63372]: DEBUG oslo_vmware.api [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 827.651738] env[63372]: value = "task-1024207" [ 827.651738] env[63372]: _type = "Task" [ 827.651738] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.660797] env[63372]: DEBUG oslo_vmware.api [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024207, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.822666] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024204, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.912835] env[63372]: DEBUG nova.network.neutron [-] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.016921] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-146a2903-f1c5-4549-8eff-d94ac6eb7c1b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.030643] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e088cb2b-6be2-431c-948b-6a814e235976 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.069275] env[63372]: DEBUG nova.compute.manager [req-d44ccfb2-a816-4a85-9734-f245cb0a44ee req-626b4ad2-8065-47d6-aaad-b2f6805fa993 service nova] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Detach interface failed, port_id=c2cf91ff-12e6-4cf7-83ac-d55b8299d8aa, reason: Instance 368101ee-e4fd-4fe4-b7c8-75c6b210189f could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 828.080152] env[63372]: DEBUG nova.compute.manager [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 828.089048] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: c825d2e2-a9cc-4c7a-b92e-039756d4121d] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 828.103386] env[63372]: DEBUG nova.scheduler.client.report [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.108923] env[63372]: DEBUG nova.virt.hardware [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.109379] env[63372]: DEBUG nova.virt.hardware [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.109379] env[63372]: DEBUG nova.virt.hardware [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.110291] env[63372]: DEBUG nova.virt.hardware [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.110291] env[63372]: DEBUG nova.virt.hardware [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.110291] env[63372]: DEBUG nova.virt.hardware [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.110291] env[63372]: DEBUG nova.virt.hardware [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.110291] env[63372]: DEBUG nova.virt.hardware [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.110505] env[63372]: DEBUG nova.virt.hardware [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.110605] env[63372]: DEBUG nova.virt.hardware [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.110805] env[63372]: DEBUG nova.virt.hardware [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.112109] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb8b28b-dbab-42a5-8cc2-f5f4ed0e6571 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.122669] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a97a7c0-00f9-4232-a8db-aa2c97150fe2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.161957] env[63372]: DEBUG oslo_vmware.api [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024207, 'name': PowerOnVM_Task, 'duration_secs': 0.495189} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.163155] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 828.163155] env[63372]: DEBUG nova.compute.manager [None req-c40f5125-1d53-4be9-84e2-153efb1fe068 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 828.163258] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe2894a-4e5a-4298-9d78-08ae01779afb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.210516] env[63372]: DEBUG nova.network.neutron [-] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.321842] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024204, 'name': CreateSnapshot_Task, 'duration_secs': 1.041115} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.322290] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 828.323048] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-348e9f6c-dfc8-4f27-b50e-9e7b64438e4d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.415373] env[63372]: INFO nova.compute.manager [-] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Took 1.39 seconds to deallocate network for instance. [ 828.556430] env[63372]: DEBUG nova.network.neutron [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Successfully updated port: 95bb06d1-3333-43b4-a667-f99370acc1d7 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 828.593012] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: e5522005-8c54-43e1-ae23-5e5ff1ef0ee9] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 828.616479] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.580s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.617007] env[63372]: DEBUG nova.compute.manager [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 828.619659] env[63372]: DEBUG oslo_concurrency.lockutils [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.254s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.619850] env[63372]: DEBUG oslo_concurrency.lockutils [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.621963] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.960s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.626019] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.626019] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.589s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.626019] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.626019] env[63372]: INFO nova.compute.manager [None req-1a882230-af30-4120-a979-6c285de524f9 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Successfully reverted task state from rebuilding on failure for instance. [ 828.629608] env[63372]: DEBUG oslo_concurrency.lockutils [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.553s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.629789] env[63372]: DEBUG oslo_concurrency.lockutils [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.631694] env[63372]: DEBUG oslo_concurrency.lockutils [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.498s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.631963] env[63372]: DEBUG oslo_concurrency.lockutils [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.633932] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.123s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.634144] env[63372]: DEBUG nova.objects.instance [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: f6675697-1529-46be-b28a-398ff3060d18] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63372) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 828.660562] env[63372]: INFO nova.scheduler.client.report [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Deleted allocations for instance f43555ef-b517-4b7d-9d2e-4787d40e201b [ 828.667086] env[63372]: INFO nova.scheduler.client.report [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Deleted allocations for instance f0c60559-c072-4b61-afe8-03d6c131b307 [ 828.671975] env[63372]: WARNING oslo_messaging._drivers.amqpdriver [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 828.687519] env[63372]: INFO nova.scheduler.client.report [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Deleted allocations for instance 5819c38e-2cf9-4d16-b28a-5f23d35c3d44 [ 828.712715] env[63372]: INFO nova.compute.manager [-] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Took 1.27 seconds to deallocate network for instance. [ 828.841037] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 828.841390] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-bc5f01ef-0560-44ab-b33e-cbd6811045f2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.851962] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 828.851962] env[63372]: value = "task-1024208" [ 828.851962] env[63372]: _type = "Task" [ 828.851962] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.861897] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024208, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.927647] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.060038] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "refresh_cache-4028d7c0-f398-4e43-9cff-5d89a14c4efa" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.060038] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "refresh_cache-4028d7c0-f398-4e43-9cff-5d89a14c4efa" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.060341] env[63372]: DEBUG nova.network.neutron [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 829.095917] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 85883c62-6d5e-4fa6-b4f0-afac984dc6bb] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 829.139572] env[63372]: DEBUG nova.compute.utils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.145067] env[63372]: DEBUG nova.compute.manager [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 829.145280] env[63372]: DEBUG nova.network.neutron [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 829.154773] env[63372]: DEBUG oslo_concurrency.lockutils [None req-946ae4d9-dfe6-46ee-a464-2258109bdb20 tempest-ServerActionsV293TestJSON-1162218784 tempest-ServerActionsV293TestJSON-1162218784-project-member] Lock "5d819bd2-f7f9-480d-83ae-fbd4d2e0d382" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.925s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.173030] env[63372]: DEBUG oslo_concurrency.lockutils [None req-62d0fde3-7276-4ebc-99f7-9f8cfe700d15 tempest-ImagesOneServerTestJSON-352743952 tempest-ImagesOneServerTestJSON-352743952-project-member] Lock "f43555ef-b517-4b7d-9d2e-4787d40e201b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.805s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.180730] env[63372]: DEBUG oslo_concurrency.lockutils [None req-985b85ca-7849-46f6-9f85-bbb689608cc6 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "f0c60559-c072-4b61-afe8-03d6c131b307" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.577s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.197195] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9d0b1540-4eb0-4d7c-8e65-908ed3470e75 tempest-ServersAdminNegativeTestJSON-1096964961 tempest-ServersAdminNegativeTestJSON-1096964961-project-member] Lock "5819c38e-2cf9-4d16-b28a-5f23d35c3d44" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.655s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.209142] env[63372]: DEBUG nova.policy [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5578d2e3e8dd498eb60b456f282d0449', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7c0777c24e2e4db083de571e3a65fa15', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 829.221859] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.242575] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "357505d0-f306-4e11-8a62-e03cfab2b7c5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.242885] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "357505d0-f306-4e11-8a62-e03cfab2b7c5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.243336] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "357505d0-f306-4e11-8a62-e03cfab2b7c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.243336] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "357505d0-f306-4e11-8a62-e03cfab2b7c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.243477] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "357505d0-f306-4e11-8a62-e03cfab2b7c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.245720] env[63372]: INFO nova.compute.manager [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Terminating instance [ 829.249021] env[63372]: DEBUG nova.compute.manager [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 829.249021] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 829.249176] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf39f2de-ab58-4706-8511-6c04805b2630 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.258875] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 829.260416] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d330bf18-58a2-4c68-841b-08fc0c241f4f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.267418] env[63372]: DEBUG oslo_vmware.api [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 829.267418] env[63372]: value = "task-1024209" [ 829.267418] env[63372]: _type = "Task" [ 829.267418] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.276470] env[63372]: DEBUG oslo_vmware.api [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024209, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.363375] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024208, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.558022] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "b182294d-2de8-4189-af7f-3e2d2c604a8b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.558022] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "b182294d-2de8-4189-af7f-3e2d2c604a8b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.558022] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "b182294d-2de8-4189-af7f-3e2d2c604a8b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.558022] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "b182294d-2de8-4189-af7f-3e2d2c604a8b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.558022] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "b182294d-2de8-4189-af7f-3e2d2c604a8b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.558419] env[63372]: INFO nova.compute.manager [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Terminating instance [ 829.560874] env[63372]: DEBUG nova.compute.manager [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 829.561064] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 829.562039] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064341ff-7d6a-4af3-bb6b-7b8d9c96eaf7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.572256] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 829.572707] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-928c8fad-ad60-41b1-9bd9-f6794f0ec37d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.581188] env[63372]: DEBUG oslo_vmware.api [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 829.581188] env[63372]: value = "task-1024210" [ 829.581188] env[63372]: _type = "Task" [ 829.581188] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.594987] env[63372]: DEBUG oslo_vmware.api [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024210, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.603840] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 44fba254-7dec-4458-9ae3-fdbbe4895de5] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 829.607622] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "93a5d948-0629-4f53-a681-858d519acfa7" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.607622] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "93a5d948-0629-4f53-a681-858d519acfa7" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.607775] env[63372]: INFO nova.compute.manager [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Rebooting instance [ 829.618021] env[63372]: DEBUG nova.network.neutron [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 829.637415] env[63372]: DEBUG nova.compute.manager [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Received event network-vif-deleted-874f14b1-6cc3-4701-8a08-35940fc9bd9b {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.637540] env[63372]: DEBUG nova.compute.manager [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Received event network-vif-plugged-95bb06d1-3333-43b4-a667-f99370acc1d7 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.637731] env[63372]: DEBUG oslo_concurrency.lockutils [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] Acquiring lock "4028d7c0-f398-4e43-9cff-5d89a14c4efa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.639192] env[63372]: DEBUG oslo_concurrency.lockutils [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] Lock "4028d7c0-f398-4e43-9cff-5d89a14c4efa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.641314] env[63372]: DEBUG oslo_concurrency.lockutils [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] Lock "4028d7c0-f398-4e43-9cff-5d89a14c4efa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.641522] env[63372]: DEBUG nova.compute.manager [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] No waiting events found dispatching network-vif-plugged-95bb06d1-3333-43b4-a667-f99370acc1d7 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 829.641764] env[63372]: WARNING nova.compute.manager [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Received unexpected event network-vif-plugged-95bb06d1-3333-43b4-a667-f99370acc1d7 for instance with vm_state building and task_state spawning. [ 829.641958] env[63372]: DEBUG nova.compute.manager [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Received event network-changed-95bb06d1-3333-43b4-a667-f99370acc1d7 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.642140] env[63372]: DEBUG nova.compute.manager [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Refreshing instance network info cache due to event network-changed-95bb06d1-3333-43b4-a667-f99370acc1d7. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 829.642323] env[63372]: DEBUG oslo_concurrency.lockutils [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] Acquiring lock "refresh_cache-4028d7c0-f398-4e43-9cff-5d89a14c4efa" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.646964] env[63372]: DEBUG nova.compute.manager [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 829.651949] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bdb150f5-de54-44fd-ada8-799c2075711a tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.656987] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.986s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.656987] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.658136] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.389s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.658337] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.660074] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.830s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.660309] env[63372]: DEBUG nova.objects.instance [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Lazy-loading 'resources' on Instance uuid 69a107d0-80c7-42e5-b514-b4273e1a3359 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 829.696442] env[63372]: INFO nova.scheduler.client.report [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Deleted allocations for instance bbba9f28-045d-41ab-8539-5b2968fe3d54 [ 829.698453] env[63372]: INFO nova.scheduler.client.report [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Deleted allocations for instance f6675697-1529-46be-b28a-398ff3060d18 [ 829.711079] env[63372]: DEBUG nova.network.neutron [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Successfully created port: 92762aef-8192-454f-95dc-1eb8d39cf9a9 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 829.780457] env[63372]: DEBUG oslo_vmware.api [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024209, 'name': PowerOffVM_Task, 'duration_secs': 0.36276} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.780796] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 829.781043] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 829.781592] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1ab1ef5-27ab-4468-b7dd-e70062117a5e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.865967] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 829.866310] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 829.866663] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Deleting the datastore file [datastore2] 357505d0-f306-4e11-8a62-e03cfab2b7c5 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 829.870506] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-343e795c-e4c8-40da-baf0-baf26dc72eb3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.872840] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024208, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.880229] env[63372]: DEBUG oslo_vmware.api [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 829.880229] env[63372]: value = "task-1024212" [ 829.880229] env[63372]: _type = "Task" [ 829.880229] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.891146] env[63372]: DEBUG oslo_vmware.api [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024212, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.929448] env[63372]: DEBUG nova.network.neutron [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Updating instance_info_cache with network_info: [{"id": "95bb06d1-3333-43b4-a667-f99370acc1d7", "address": "fa:16:3e:6e:68:90", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95bb06d1-33", "ovs_interfaceid": "95bb06d1-3333-43b4-a667-f99370acc1d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.981042] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d21b3e-b8d2-3614-b4cf-1ac8c9255696/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 829.981202] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036d8c4f-36f7-4b6a-a624-075e2305afd8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.993670] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d21b3e-b8d2-3614-b4cf-1ac8c9255696/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 829.993670] env[63372]: ERROR oslo_vmware.rw_handles [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d21b3e-b8d2-3614-b4cf-1ac8c9255696/disk-0.vmdk due to incomplete transfer. [ 829.993670] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-14e5a556-05aa-4b5d-b1fe-857d847f1608 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.006024] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d21b3e-b8d2-3614-b4cf-1ac8c9255696/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 830.006024] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Uploaded image 77e136a4-4a74-4ff7-8ff7-0eb4fa83bb02 to the Glance image server {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 830.008370] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 830.008991] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b8ba2fe1-8302-4e2b-8c26-d07046d1f26d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.016489] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 830.016489] env[63372]: value = "task-1024213" [ 830.016489] env[63372]: _type = "Task" [ 830.016489] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.027754] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024213, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.094153] env[63372]: DEBUG oslo_vmware.api [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024210, 'name': PowerOffVM_Task, 'duration_secs': 0.215544} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.094663] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 830.094663] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 830.094913] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-06a77929-ca64-4eb2-8c10-e4acef57aafa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.107836] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 276c3490-240a-499b-ba64-4b02510b169c] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 830.126593] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.126780] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.126959] env[63372]: DEBUG nova.network.neutron [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 830.168538] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 830.168538] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 830.168538] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Deleting the datastore file [datastore1] b182294d-2de8-4189-af7f-3e2d2c604a8b {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 830.168841] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8aec6ee8-6514-4acc-a16b-7a5f5f4bf744 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.176598] env[63372]: DEBUG oslo_vmware.api [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for the task: (returnval){ [ 830.176598] env[63372]: value = "task-1024215" [ 830.176598] env[63372]: _type = "Task" [ 830.176598] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.191096] env[63372]: DEBUG oslo_vmware.api [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024215, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.210851] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f8ebe864-4db1-4a6f-a21a-294507816d5b tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "bbba9f28-045d-41ab-8539-5b2968fe3d54" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.540s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.213350] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5cbcc352-b8a3-4964-90b3-cbbdab814a6e tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "f6675697-1529-46be-b28a-398ff3060d18" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.427s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.374454] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024208, 'name': CloneVM_Task} progress is 95%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.395937] env[63372]: DEBUG oslo_vmware.api [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024212, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.423614} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.401443] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 830.401443] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 830.401443] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 830.401443] env[63372]: INFO nova.compute.manager [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Took 1.15 seconds to destroy the instance on the hypervisor. [ 830.401443] env[63372]: DEBUG oslo.service.loopingcall [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.401443] env[63372]: DEBUG nova.compute.manager [-] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 830.401443] env[63372]: DEBUG nova.network.neutron [-] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 830.433580] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "refresh_cache-4028d7c0-f398-4e43-9cff-5d89a14c4efa" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.434108] env[63372]: DEBUG nova.compute.manager [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Instance network_info: |[{"id": "95bb06d1-3333-43b4-a667-f99370acc1d7", "address": "fa:16:3e:6e:68:90", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95bb06d1-33", "ovs_interfaceid": "95bb06d1-3333-43b4-a667-f99370acc1d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 830.434507] env[63372]: DEBUG oslo_concurrency.lockutils [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] Acquired lock "refresh_cache-4028d7c0-f398-4e43-9cff-5d89a14c4efa" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.434752] env[63372]: DEBUG nova.network.neutron [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Refreshing network info cache for port 95bb06d1-3333-43b4-a667-f99370acc1d7 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 830.439110] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:68:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60badc2d-69d2-467d-a92e-98511f5cb0b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '95bb06d1-3333-43b4-a667-f99370acc1d7', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.444848] env[63372]: DEBUG oslo.service.loopingcall [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.454548] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 830.454548] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eb4fa617-769f-4933-af66-b2f37fa895de {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.477041] env[63372]: DEBUG oslo_concurrency.lockutils [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "3f66eda6-f5e9-4527-9711-849a01702580" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.477318] env[63372]: DEBUG oslo_concurrency.lockutils [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "3f66eda6-f5e9-4527-9711-849a01702580" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.477527] env[63372]: DEBUG oslo_concurrency.lockutils [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "3f66eda6-f5e9-4527-9711-849a01702580-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.477714] env[63372]: DEBUG oslo_concurrency.lockutils [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "3f66eda6-f5e9-4527-9711-849a01702580-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.477913] env[63372]: DEBUG oslo_concurrency.lockutils [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "3f66eda6-f5e9-4527-9711-849a01702580-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.481166] env[63372]: INFO nova.compute.manager [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Terminating instance [ 830.482969] env[63372]: DEBUG oslo_concurrency.lockutils [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "refresh_cache-3f66eda6-f5e9-4527-9711-849a01702580" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.483205] env[63372]: DEBUG oslo_concurrency.lockutils [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquired lock "refresh_cache-3f66eda6-f5e9-4527-9711-849a01702580" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.483405] env[63372]: DEBUG nova.network.neutron [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 830.485781] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.485781] env[63372]: value = "task-1024216" [ 830.485781] env[63372]: _type = "Task" [ 830.485781] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.499144] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024216, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.528631] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024213, 'name': Destroy_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.616151] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 5adf6952-e019-4ea0-be91-8fe5fb83d53e] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 830.633865] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0368b42-ea20-40b7-9233-359c00fb66ce {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.644839] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c06b1cf-c430-4b82-aabb-d14dc47552bb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.677827] env[63372]: DEBUG nova.compute.manager [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 830.684306] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751abbcd-7f6b-4d37-8494-09186f7ae8bf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.698256] env[63372]: DEBUG oslo_vmware.api [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Task: {'id': task-1024215, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.342391} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.701027] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 830.702027] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 830.702027] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 830.702027] env[63372]: INFO nova.compute.manager [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 830.702027] env[63372]: DEBUG oslo.service.loopingcall [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.702213] env[63372]: DEBUG nova.compute.manager [-] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 830.702919] env[63372]: DEBUG nova.network.neutron [-] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 830.705026] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121996cf-08c5-4b48-b732-7837e33537f4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.726075] env[63372]: DEBUG nova.compute.provider_tree [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.731065] env[63372]: DEBUG nova.virt.hardware [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 830.731065] env[63372]: DEBUG nova.virt.hardware [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 830.731065] env[63372]: DEBUG nova.virt.hardware [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.731245] env[63372]: DEBUG nova.virt.hardware [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 830.731401] env[63372]: DEBUG nova.virt.hardware [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.731552] env[63372]: DEBUG nova.virt.hardware [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 830.731763] env[63372]: DEBUG nova.virt.hardware [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 830.731924] env[63372]: DEBUG nova.virt.hardware [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 830.732133] env[63372]: DEBUG nova.virt.hardware [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 830.733096] env[63372]: DEBUG nova.virt.hardware [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 830.733096] env[63372]: DEBUG nova.virt.hardware [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.735412] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33632e5-6e68-4a6b-a0c2-ab7f113e97ab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.745608] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78c026c-7b25-46ca-a710-33d019ca203f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.868369] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024208, 'name': CloneVM_Task, 'duration_secs': 1.703369} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.868369] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Created linked-clone VM from snapshot [ 830.869250] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279e479b-6ab3-40da-8d31-440f6af49aae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.877324] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Uploading image 5105490f-194c-4b4c-9151-ee2bf01d1663 {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 830.888325] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 830.888597] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-be2bccf5-a2d0-40dd-be85-d31d2d874723 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.902069] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 830.902069] env[63372]: value = "task-1024217" [ 830.902069] env[63372]: _type = "Task" [ 830.902069] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.911817] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024217, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.929070] env[63372]: DEBUG nova.network.neutron [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance_info_cache with network_info: [{"id": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "address": "fa:16:3e:3f:99:77", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c4c74ba-ad", "ovs_interfaceid": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.999238] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024216, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.012209] env[63372]: DEBUG nova.network.neutron [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 831.028417] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024213, 'name': Destroy_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.078121] env[63372]: DEBUG nova.network.neutron [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.121697] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 9404a0f0-a3e5-43ee-bc20-e9566cfe5a60] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 831.239992] env[63372]: DEBUG nova.scheduler.client.report [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.244258] env[63372]: DEBUG nova.network.neutron [-] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.416989] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024217, 'name': Destroy_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.426770] env[63372]: DEBUG nova.network.neutron [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Updated VIF entry in instance network info cache for port 95bb06d1-3333-43b4-a667-f99370acc1d7. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 831.426944] env[63372]: DEBUG nova.network.neutron [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Updating instance_info_cache with network_info: [{"id": "95bb06d1-3333-43b4-a667-f99370acc1d7", "address": "fa:16:3e:6e:68:90", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95bb06d1-33", "ovs_interfaceid": "95bb06d1-3333-43b4-a667-f99370acc1d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.431538] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.434373] env[63372]: DEBUG nova.compute.manager [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 831.435639] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3d7542-5bdf-479c-a30e-2c744472946a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.503757] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024216, 'name': CreateVM_Task, 'duration_secs': 0.927495} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.504757] env[63372]: DEBUG nova.network.neutron [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Successfully updated port: 92762aef-8192-454f-95dc-1eb8d39cf9a9 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 831.506079] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 831.506449] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.506621] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.507069] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.510473] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bdd3e50-ceb5-4b73-88a9-8fa3dea38415 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.515732] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 831.515732] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520171c0-da36-3391-8181-37c60d5065e9" [ 831.515732] env[63372]: _type = "Task" [ 831.515732] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.532919] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520171c0-da36-3391-8181-37c60d5065e9, 'name': SearchDatastore_Task, 'duration_secs': 0.012788} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.539053] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.539053] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.539053] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.539053] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.539053] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 831.539053] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024213, 'name': Destroy_Task, 'duration_secs': 1.109376} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.539053] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2aa7d818-4663-4f2d-8349-1eb6644d1eff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.539713] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Destroyed the VM [ 831.539943] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 831.544300] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-516534dc-d81c-481c-ba4e-837e61398738 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.553875] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 831.553875] env[63372]: value = "task-1024218" [ 831.553875] env[63372]: _type = "Task" [ 831.553875] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.555261] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 831.555477] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 831.559827] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aeef2ef6-7c96-4cbf-af2d-695518ca9814 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.571023] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024218, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.572682] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 831.572682] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5287f996-787a-7971-9dc0-63f6628b6379" [ 831.572682] env[63372]: _type = "Task" [ 831.572682] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.580057] env[63372]: DEBUG oslo_concurrency.lockutils [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Releasing lock "refresh_cache-3f66eda6-f5e9-4527-9711-849a01702580" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.580362] env[63372]: DEBUG nova.compute.manager [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 831.580562] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 831.585019] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d1cf40-8026-4cf6-92fc-0dd593219773 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.587875] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5287f996-787a-7971-9dc0-63f6628b6379, 'name': SearchDatastore_Task, 'duration_secs': 0.011052} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.589144] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1af2995-7795-4dff-8423-c5e48f0f3d9d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.594300] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 831.594926] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1202eb74-3cc4-48a1-bba9-53dd199dd3c2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.600075] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 831.600075] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]524f2194-b90a-cb95-962e-65fa2fbed237" [ 831.600075] env[63372]: _type = "Task" [ 831.600075] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.605890] env[63372]: DEBUG oslo_vmware.api [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 831.605890] env[63372]: value = "task-1024219" [ 831.605890] env[63372]: _type = "Task" [ 831.605890] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.609300] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524f2194-b90a-cb95-962e-65fa2fbed237, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.621866] env[63372]: DEBUG oslo_vmware.api [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024219, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.623727] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 779e3559-37a5-49b2-b17a-bf95d0d8d23c] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 831.747546] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.087s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.750125] env[63372]: INFO nova.compute.manager [-] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Took 1.35 seconds to deallocate network for instance. [ 831.750125] env[63372]: DEBUG oslo_concurrency.lockutils [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.795s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.750446] env[63372]: DEBUG nova.objects.instance [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Lazy-loading 'resources' on Instance uuid d50919ac-3a0b-46ac-a837-ca3e6ad05173 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.775449] env[63372]: INFO nova.scheduler.client.report [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Deleted allocations for instance 69a107d0-80c7-42e5-b514-b4273e1a3359 [ 831.915766] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024217, 'name': Destroy_Task, 'duration_secs': 0.742123} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.917156] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Destroyed the VM [ 831.917448] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 831.917726] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a2dde435-caff-442c-aec3-82225339399d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.927098] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 831.927098] env[63372]: value = "task-1024220" [ 831.927098] env[63372]: _type = "Task" [ 831.927098] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.931026] env[63372]: DEBUG oslo_concurrency.lockutils [req-15b7789b-fe3b-4866-87fe-2994cc5cc8ba req-c23e97f1-b109-4824-9f77-3c742bfe8c43 service nova] Releasing lock "refresh_cache-4028d7c0-f398-4e43-9cff-5d89a14c4efa" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.942246] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024220, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.956243] env[63372]: DEBUG nova.compute.manager [req-a672774f-68d4-49d0-94c7-d7ab53d5a13d req-d8a63a49-ed44-44fe-957f-cac2fdc8123b service nova] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Received event network-vif-deleted-c56eef25-45d9-459d-b839-c069b18a613c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.956339] env[63372]: INFO nova.compute.manager [req-a672774f-68d4-49d0-94c7-d7ab53d5a13d req-d8a63a49-ed44-44fe-957f-cac2fdc8123b service nova] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Neutron deleted interface c56eef25-45d9-459d-b839-c069b18a613c; detaching it from the instance and deleting it from the info cache [ 831.956507] env[63372]: DEBUG nova.network.neutron [req-a672774f-68d4-49d0-94c7-d7ab53d5a13d req-d8a63a49-ed44-44fe-957f-cac2fdc8123b service nova] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.981878] env[63372]: DEBUG nova.compute.manager [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Received event network-vif-deleted-bf61306f-1264-4480-b485-299826369f7d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.982140] env[63372]: DEBUG nova.compute.manager [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Received event network-vif-plugged-92762aef-8192-454f-95dc-1eb8d39cf9a9 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.982357] env[63372]: DEBUG oslo_concurrency.lockutils [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] Acquiring lock "a3d5b93e-1b76-48e7-bb3e-2423e7c4119a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.982603] env[63372]: DEBUG oslo_concurrency.lockutils [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] Lock "a3d5b93e-1b76-48e7-bb3e-2423e7c4119a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.982795] env[63372]: DEBUG oslo_concurrency.lockutils [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] Lock "a3d5b93e-1b76-48e7-bb3e-2423e7c4119a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.982974] env[63372]: DEBUG nova.compute.manager [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] No waiting events found dispatching network-vif-plugged-92762aef-8192-454f-95dc-1eb8d39cf9a9 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 831.983156] env[63372]: WARNING nova.compute.manager [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Received unexpected event network-vif-plugged-92762aef-8192-454f-95dc-1eb8d39cf9a9 for instance with vm_state building and task_state spawning. [ 831.983320] env[63372]: DEBUG nova.compute.manager [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Received event network-changed-92762aef-8192-454f-95dc-1eb8d39cf9a9 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.983474] env[63372]: DEBUG nova.compute.manager [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Refreshing instance network info cache due to event network-changed-92762aef-8192-454f-95dc-1eb8d39cf9a9. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.983663] env[63372]: DEBUG oslo_concurrency.lockutils [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] Acquiring lock "refresh_cache-a3d5b93e-1b76-48e7-bb3e-2423e7c4119a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.983801] env[63372]: DEBUG oslo_concurrency.lockutils [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] Acquired lock "refresh_cache-a3d5b93e-1b76-48e7-bb3e-2423e7c4119a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.984010] env[63372]: DEBUG nova.network.neutron [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Refreshing network info cache for port 92762aef-8192-454f-95dc-1eb8d39cf9a9 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 832.010767] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "refresh_cache-a3d5b93e-1b76-48e7-bb3e-2423e7c4119a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.067079] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024218, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.101097] env[63372]: DEBUG nova.network.neutron [-] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.112302] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524f2194-b90a-cb95-962e-65fa2fbed237, 'name': SearchDatastore_Task, 'duration_secs': 0.011803} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.115373] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.115878] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 4028d7c0-f398-4e43-9cff-5d89a14c4efa/4028d7c0-f398-4e43-9cff-5d89a14c4efa.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 832.119161] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8641b8b4-39af-4907-82ae-382a6d341002 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.124803] env[63372]: DEBUG oslo_vmware.api [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024219, 'name': PowerOffVM_Task, 'duration_secs': 0.131131} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.126339] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 832.126424] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 832.126796] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 832.126796] env[63372]: value = "task-1024221" [ 832.126796] env[63372]: _type = "Task" [ 832.126796] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.127012] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45497517-964c-4f36-9544-99a42417d195 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.128899] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 6d4502b7-b9df-4fef-8400-9ff07b820e2b] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 832.140764] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024221, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.164758] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 832.164990] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 832.165193] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Deleting the datastore file [datastore1] 3f66eda6-f5e9-4527-9711-849a01702580 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.165755] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a23c0e81-37f8-42e3-92f8-a2594705e0d5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.174694] env[63372]: DEBUG oslo_vmware.api [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for the task: (returnval){ [ 832.174694] env[63372]: value = "task-1024223" [ 832.174694] env[63372]: _type = "Task" [ 832.174694] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.185850] env[63372]: DEBUG oslo_vmware.api [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024223, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.260620] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.286653] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4b008ca7-4464-4d66-b3ec-9eed58e315b4 tempest-ServersTestBootFromVolume-1129708801 tempest-ServersTestBootFromVolume-1129708801-project-member] Lock "69a107d0-80c7-42e5-b514-b4273e1a3359" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.539s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.442946] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024220, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.463230] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b885f5e-81c9-49a7-bf62-7a7058f09a88 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.466331] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2335b9a3-e5ab-4990-a986-1cf594f70e25 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.473779] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Doing hard reboot of VM {{(pid=63372) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 832.474075] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-0752073d-dfa6-4069-a0c3-e97d17e33b0c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.482208] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767b77f2-16e4-4d49-87f9-ebe9197a139d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.497187] env[63372]: DEBUG oslo_vmware.api [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 832.497187] env[63372]: value = "task-1024224" [ 832.497187] env[63372]: _type = "Task" [ 832.497187] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.507785] env[63372]: DEBUG oslo_vmware.api [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024224, 'name': ResetVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.529049] env[63372]: DEBUG nova.compute.manager [req-a672774f-68d4-49d0-94c7-d7ab53d5a13d req-d8a63a49-ed44-44fe-957f-cac2fdc8123b service nova] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Detach interface failed, port_id=c56eef25-45d9-459d-b839-c069b18a613c, reason: Instance b182294d-2de8-4189-af7f-3e2d2c604a8b could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 832.564235] env[63372]: DEBUG nova.network.neutron [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 832.574033] env[63372]: DEBUG oslo_vmware.api [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024218, 'name': RemoveSnapshot_Task, 'duration_secs': 0.547947} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.574033] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 832.574033] env[63372]: INFO nova.compute.manager [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Took 18.45 seconds to snapshot the instance on the hypervisor. [ 832.612224] env[63372]: INFO nova.compute.manager [-] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Took 1.90 seconds to deallocate network for instance. [ 832.639397] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 715408b8-0e76-4b61-a342-b168377cb288] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 832.649817] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024221, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.685492] env[63372]: DEBUG oslo_vmware.api [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Task: {'id': task-1024223, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161479} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.685865] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 832.686286] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 832.686798] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 832.687123] env[63372]: INFO nova.compute.manager [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Took 1.11 seconds to destroy the instance on the hypervisor. [ 832.687561] env[63372]: DEBUG oslo.service.loopingcall [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.687981] env[63372]: DEBUG nova.compute.manager [-] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 832.688134] env[63372]: DEBUG nova.network.neutron [-] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 832.713523] env[63372]: DEBUG nova.network.neutron [-] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 832.722436] env[63372]: DEBUG nova.network.neutron [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.727520] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef5e9ce2-7f2f-4fdb-aeb3-adef7a2fba69 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.736295] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2bc2da5-4300-4408-9929-6a25012c2f8e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.771213] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad541aa1-9a85-4f90-8673-5d49034653a6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.780130] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45cb4cc-246d-4d19-a5ed-e5edb598fab7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.794552] env[63372]: DEBUG nova.compute.provider_tree [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.942138] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024220, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.009021] env[63372]: DEBUG oslo_vmware.api [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024224, 'name': ResetVM_Task, 'duration_secs': 0.127433} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.009021] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Did hard reboot of VM {{(pid=63372) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 833.009021] env[63372]: DEBUG nova.compute.manager [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.009543] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0cd1af4-505e-44a5-b9f6-bf1510d97ccd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.117610] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.124243] env[63372]: DEBUG nova.compute.manager [None req-c29866d3-3c05-4618-8c48-e3399eaaaeda tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Found 2 images (rotation: 2) {{(pid=63372) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 833.142629] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024221, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.569007} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.143097] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 8986e782-4d63-4491-8100-4341a3149812] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 833.144877] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 4028d7c0-f398-4e43-9cff-5d89a14c4efa/4028d7c0-f398-4e43-9cff-5d89a14c4efa.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 833.145117] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 833.146046] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f1b771a0-03e1-4e77-adb1-1cd80ef38455 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.154911] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 833.154911] env[63372]: value = "task-1024225" [ 833.154911] env[63372]: _type = "Task" [ 833.154911] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.170142] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024225, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.216460] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Acquiring lock "cf673ac1-2c7d-468b-83ec-c723d5182457" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.216792] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Lock "cf673ac1-2c7d-468b-83ec-c723d5182457" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.217796] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Acquiring lock "cf673ac1-2c7d-468b-83ec-c723d5182457-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.217796] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Lock "cf673ac1-2c7d-468b-83ec-c723d5182457-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.217796] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Lock "cf673ac1-2c7d-468b-83ec-c723d5182457-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.221297] env[63372]: DEBUG nova.network.neutron [-] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.225685] env[63372]: INFO nova.compute.manager [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Terminating instance [ 833.225685] env[63372]: DEBUG oslo_concurrency.lockutils [req-e6bb9cd0-564e-493b-a028-1d31a978ca47 req-6b65bcb1-02b8-409d-84f5-c477d8f329b5 service nova] Releasing lock "refresh_cache-a3d5b93e-1b76-48e7-bb3e-2423e7c4119a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.225685] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquired lock "refresh_cache-a3d5b93e-1b76-48e7-bb3e-2423e7c4119a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.225923] env[63372]: DEBUG nova.network.neutron [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 833.227544] env[63372]: DEBUG nova.compute.manager [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 833.229624] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 833.229624] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f3bb42-2366-44b4-9578-817b4e823847 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.238112] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 833.238398] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7545d216-1053-4e90-90a9-15f7ea275bfe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.254962] env[63372]: DEBUG oslo_vmware.api [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Waiting for the task: (returnval){ [ 833.254962] env[63372]: value = "task-1024226" [ 833.254962] env[63372]: _type = "Task" [ 833.254962] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.269686] env[63372]: DEBUG oslo_vmware.api [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1024226, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.302176] env[63372]: DEBUG nova.scheduler.client.report [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.440700] env[63372]: DEBUG oslo_vmware.api [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024220, 'name': RemoveSnapshot_Task, 'duration_secs': 1.122204} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.441016] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 833.523076] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0079053c-20b1-49b7-b3e5-d4b618de343a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "93a5d948-0629-4f53-a681-858d519acfa7" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.915s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.646885] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 833.647097] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Cleaning up deleted instances with incomplete migration {{(pid=63372) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 833.666106] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024225, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065111} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.666382] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 833.667350] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff397fe2-5354-4cba-b43b-1502a4c4faa7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.691149] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 4028d7c0-f398-4e43-9cff-5d89a14c4efa/4028d7c0-f398-4e43-9cff-5d89a14c4efa.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.691470] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8bbdadc3-e7d7-497a-87fc-a407f4215ded {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.713407] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 833.713407] env[63372]: value = "task-1024227" [ 833.713407] env[63372]: _type = "Task" [ 833.713407] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.722696] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024227, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.725937] env[63372]: INFO nova.compute.manager [-] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Took 1.04 seconds to deallocate network for instance. [ 833.765451] env[63372]: DEBUG oslo_vmware.api [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1024226, 'name': PowerOffVM_Task, 'duration_secs': 0.333061} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.766321] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 833.766486] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 833.766778] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7ae0913d-fba8-49ad-945f-35c5051d6d61 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.769200] env[63372]: DEBUG nova.network.neutron [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 833.811079] env[63372]: DEBUG oslo_concurrency.lockutils [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.061s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.813390] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.135s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.815298] env[63372]: INFO nova.compute.claims [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 833.843499] env[63372]: INFO nova.scheduler.client.report [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Deleted allocations for instance d50919ac-3a0b-46ac-a837-ca3e6ad05173 [ 833.851236] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 833.851649] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 833.851649] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Deleting the datastore file [datastore2] cf673ac1-2c7d-468b-83ec-c723d5182457 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 833.851754] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7824ee51-2402-4205-a07e-bdf7c78f6b17 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.861887] env[63372]: DEBUG oslo_vmware.api [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Waiting for the task: (returnval){ [ 833.861887] env[63372]: value = "task-1024229" [ 833.861887] env[63372]: _type = "Task" [ 833.861887] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.875031] env[63372]: DEBUG oslo_vmware.api [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1024229, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.947200] env[63372]: WARNING nova.compute.manager [None req-a9200ff0-f71a-45a6-b58e-63ab6b5218a2 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Image not found during snapshot: nova.exception.ImageNotFound: Image 5105490f-194c-4b4c-9151-ee2bf01d1663 could not be found. [ 834.023634] env[63372]: DEBUG nova.network.neutron [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Updating instance_info_cache with network_info: [{"id": "92762aef-8192-454f-95dc-1eb8d39cf9a9", "address": "fa:16:3e:08:ce:c1", "network": {"id": "d4022663-3d99-4ea6-b2c8-7bf206d3a801", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2038051617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c0777c24e2e4db083de571e3a65fa15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ee018eb-75be-4037-a80a-07034d4eae35", "external-id": "nsx-vlan-transportzone-8", "segmentation_id": 8, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92762aef-81", "ovs_interfaceid": "92762aef-8192-454f-95dc-1eb8d39cf9a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.149646] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 834.227721] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024227, 'name': ReconfigVM_Task, 'duration_secs': 0.318667} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.228080] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 4028d7c0-f398-4e43-9cff-5d89a14c4efa/4028d7c0-f398-4e43-9cff-5d89a14c4efa.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 834.228813] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-423a17cd-c3bd-4b8a-98fe-0e57d9cd4eb9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.233045] env[63372]: DEBUG oslo_concurrency.lockutils [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.239816] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 834.239816] env[63372]: value = "task-1024230" [ 834.239816] env[63372]: _type = "Task" [ 834.239816] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.256404] env[63372]: DEBUG nova.compute.manager [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 834.256761] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024230, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.257588] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d96bd9c-59dc-4fbf-bad0-7e6ccb873b65 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.353998] env[63372]: DEBUG oslo_concurrency.lockutils [None req-438c1e2e-dfa3-4c63-8431-e0550180b90b tempest-AttachInterfacesUnderV243Test-648476154 tempest-AttachInterfacesUnderV243Test-648476154-project-member] Lock "d50919ac-3a0b-46ac-a837-ca3e6ad05173" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.179s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.376349] env[63372]: DEBUG oslo_vmware.api [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Task: {'id': task-1024229, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.398546} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.376637] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.376828] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 834.377017] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 834.377248] env[63372]: INFO nova.compute.manager [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Took 1.15 seconds to destroy the instance on the hypervisor. [ 834.377521] env[63372]: DEBUG oslo.service.loopingcall [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.377733] env[63372]: DEBUG nova.compute.manager [-] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.377821] env[63372]: DEBUG nova.network.neutron [-] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 834.531287] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Releasing lock "refresh_cache-a3d5b93e-1b76-48e7-bb3e-2423e7c4119a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.531671] env[63372]: DEBUG nova.compute.manager [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Instance network_info: |[{"id": "92762aef-8192-454f-95dc-1eb8d39cf9a9", "address": "fa:16:3e:08:ce:c1", "network": {"id": "d4022663-3d99-4ea6-b2c8-7bf206d3a801", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2038051617-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c0777c24e2e4db083de571e3a65fa15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ee018eb-75be-4037-a80a-07034d4eae35", "external-id": "nsx-vlan-transportzone-8", "segmentation_id": 8, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92762aef-81", "ovs_interfaceid": "92762aef-8192-454f-95dc-1eb8d39cf9a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 834.532168] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:ce:c1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ee018eb-75be-4037-a80a-07034d4eae35', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '92762aef-8192-454f-95dc-1eb8d39cf9a9', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 834.546472] env[63372]: DEBUG oslo.service.loopingcall [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.546472] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 834.546472] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a5844ea-feb5-423f-84be-8b31a80945af {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.573951] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 834.573951] env[63372]: value = "task-1024231" [ 834.573951] env[63372]: _type = "Task" [ 834.573951] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.592652] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024231, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.686599] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "13bde897-8446-42a2-b02d-2f5b48e6f432" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.686869] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "13bde897-8446-42a2-b02d-2f5b48e6f432" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.755020] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024230, 'name': Rename_Task, 'duration_secs': 0.184334} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.755020] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 834.755020] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6dfa200a-f357-4616-aa67-aa0db8258380 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.767057] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 834.767057] env[63372]: value = "task-1024232" [ 834.767057] env[63372]: _type = "Task" [ 834.767057] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.775736] env[63372]: INFO nova.compute.manager [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] instance snapshotting [ 834.776700] env[63372]: DEBUG nova.objects.instance [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'flavor' on Instance uuid 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 834.800039] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024232, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.092310] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024231, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.183785] env[63372]: DEBUG nova.compute.manager [req-6accf3ff-937d-4f97-ba02-398caba088f4 req-ebd94354-31cf-4d91-8755-2783f54c77fb service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Received event network-vif-deleted-e56add8e-4342-4ab4-b2f2-c03f257f3d3a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.183933] env[63372]: INFO nova.compute.manager [req-6accf3ff-937d-4f97-ba02-398caba088f4 req-ebd94354-31cf-4d91-8755-2783f54c77fb service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Neutron deleted interface e56add8e-4342-4ab4-b2f2-c03f257f3d3a; detaching it from the instance and deleting it from the info cache [ 835.187231] env[63372]: DEBUG nova.network.neutron [req-6accf3ff-937d-4f97-ba02-398caba088f4 req-ebd94354-31cf-4d91-8755-2783f54c77fb service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.190976] env[63372]: DEBUG nova.compute.manager [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 835.228904] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77697519-d195-4620-9dcf-4eb9ea0b484d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.238279] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ee641b-d617-4bc9-965b-a1012231ebbb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.285033] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "89301344-84a9-4d13-aae7-99943d0a478e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.285307] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "89301344-84a9-4d13-aae7-99943d0a478e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.285601] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "89301344-84a9-4d13-aae7-99943d0a478e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.285661] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "89301344-84a9-4d13-aae7-99943d0a478e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.285811] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "89301344-84a9-4d13-aae7-99943d0a478e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.293523] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a08a16-7b01-45f3-bcd0-2cbb2e489ecf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.296911] env[63372]: INFO nova.compute.manager [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Terminating instance [ 835.299199] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1d645a-91f7-4595-9db1-5fed69af8cf4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.303419] env[63372]: DEBUG nova.compute.manager [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 835.303419] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 835.303999] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b24fc08-e07a-4f04-befc-ae2e6ff39b4d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.313736] env[63372]: DEBUG oslo_vmware.api [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024232, 'name': PowerOnVM_Task, 'duration_secs': 0.51209} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.314487] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 835.314702] env[63372]: INFO nova.compute.manager [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Took 7.23 seconds to spawn the instance on the hypervisor. [ 835.314888] env[63372]: DEBUG nova.compute.manager [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 835.316166] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e9d1d29-d9b8-4950-adac-760fa80b7377 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.341218] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16d8bec-7af7-4b3f-8138-afa79f7f148f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.344206] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 835.345340] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820a1936-7dfd-47f4-abbf-96c3bb1cc170 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.349165] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-68598058-b7da-416f-80b6-8a5e3f8e554c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.359124] env[63372]: DEBUG nova.compute.provider_tree [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.371118] env[63372]: DEBUG oslo_vmware.api [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 835.371118] env[63372]: value = "task-1024233" [ 835.371118] env[63372]: _type = "Task" [ 835.371118] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.382850] env[63372]: DEBUG oslo_vmware.api [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024233, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.568425] env[63372]: DEBUG nova.network.neutron [-] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.591901] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024231, 'name': CreateVM_Task, 'duration_secs': 0.562052} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.591901] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 835.591901] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.592552] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.592552] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 835.592862] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b6329b2-5d78-4e4a-82d0-97f63a3b717a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.600064] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 835.600064] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]522db497-98e3-9c51-08a5-8117607144b1" [ 835.600064] env[63372]: _type = "Task" [ 835.600064] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.609531] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522db497-98e3-9c51-08a5-8117607144b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.689603] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-837266b9-9421-4354-9813-8251ccfb7e21 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.704862] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c6241cb-677b-400e-9794-d70ea2127395 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.749659] env[63372]: DEBUG nova.compute.manager [req-6accf3ff-937d-4f97-ba02-398caba088f4 req-ebd94354-31cf-4d91-8755-2783f54c77fb service nova] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Detach interface failed, port_id=e56add8e-4342-4ab4-b2f2-c03f257f3d3a, reason: Instance cf673ac1-2c7d-468b-83ec-c723d5182457 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 835.750945] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.871097] env[63372]: DEBUG nova.scheduler.client.report [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 835.878664] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 835.888890] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-686135ff-d0e0-43a1-8615-d2b8eff27066 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.892616] env[63372]: INFO nova.compute.manager [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Took 40.83 seconds to build instance. [ 835.901276] env[63372]: DEBUG oslo_vmware.api [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024233, 'name': PowerOffVM_Task, 'duration_secs': 0.23707} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.902822] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 835.902982] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 835.903397] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 835.903397] env[63372]: value = "task-1024234" [ 835.903397] env[63372]: _type = "Task" [ 835.903397] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.904200] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-13fb958a-8cbf-47fa-9cfa-08847ceb5fdc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.916459] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024234, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.984505] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 835.984652] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 835.985764] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Deleting the datastore file [datastore2] 89301344-84a9-4d13-aae7-99943d0a478e {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.985764] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5a025ba7-62bf-4f39-a8de-87faadd21a24 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.992760] env[63372]: DEBUG oslo_vmware.api [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 835.992760] env[63372]: value = "task-1024236" [ 835.992760] env[63372]: _type = "Task" [ 835.992760] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.002662] env[63372]: DEBUG oslo_vmware.api [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024236, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.077614] env[63372]: INFO nova.compute.manager [-] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Took 1.70 seconds to deallocate network for instance. [ 836.113040] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522db497-98e3-9c51-08a5-8117607144b1, 'name': SearchDatastore_Task, 'duration_secs': 0.031181} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.113040] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.113040] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.113040] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.113040] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.113040] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.113040] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ffd40be1-2ce1-478d-80e0-ddf5d6775c3d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.122358] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.122531] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 836.123289] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53f004b4-2e42-4b67-b8d1-8bbbdf6e8b6c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.130840] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 836.130840] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525ba741-3974-8460-46dd-bbb33b6517a9" [ 836.130840] env[63372]: _type = "Task" [ 836.130840] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.139172] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525ba741-3974-8460-46dd-bbb33b6517a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.382259] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.566s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.382259] env[63372]: DEBUG nova.compute.manager [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 836.383104] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.178s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.384682] env[63372]: INFO nova.compute.claims [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.395932] env[63372]: DEBUG oslo_concurrency.lockutils [None req-da2acf3d-284c-4459-9ca4-7d6f39b5094c tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "4028d7c0-f398-4e43-9cff-5d89a14c4efa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.347s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.417969] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024234, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.505698] env[63372]: DEBUG oslo_vmware.api [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024236, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.248731} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.506459] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 836.506459] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 836.506655] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 836.506889] env[63372]: INFO nova.compute.manager [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Took 1.20 seconds to destroy the instance on the hypervisor. [ 836.507233] env[63372]: DEBUG oslo.service.loopingcall [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.507507] env[63372]: DEBUG nova.compute.manager [-] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 836.507639] env[63372]: DEBUG nova.network.neutron [-] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 836.588583] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.643865] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525ba741-3974-8460-46dd-bbb33b6517a9, 'name': SearchDatastore_Task, 'duration_secs': 0.015056} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.644963] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e0d6b53-8f7e-4c6c-96f8-a3fef4f5ae93 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.651506] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 836.651506] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5231eaeb-d0e2-2bc9-10b6-650cbab253c7" [ 836.651506] env[63372]: _type = "Task" [ 836.651506] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.662323] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5231eaeb-d0e2-2bc9-10b6-650cbab253c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.890602] env[63372]: DEBUG nova.compute.utils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 836.902227] env[63372]: DEBUG nova.compute.manager [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 836.902326] env[63372]: DEBUG nova.network.neutron [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 836.918804] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024234, 'name': CreateSnapshot_Task, 'duration_secs': 0.672653} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.918948] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 836.919914] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7cebe2-4027-4de7-a660-a2a5daa4e213 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.981356] env[63372]: DEBUG nova.policy [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b43b9a0ba7094627b100ee4460eabb67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '56e781c01b95446b91c6beae143d1be4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 837.012939] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquiring lock "da8cf1ff-d977-4fd1-9239-043f07bed875" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.013207] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Lock "da8cf1ff-d977-4fd1-9239-043f07bed875" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.166950] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5231eaeb-d0e2-2bc9-10b6-650cbab253c7, 'name': SearchDatastore_Task, 'duration_secs': 0.011408} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.167460] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.167796] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] a3d5b93e-1b76-48e7-bb3e-2423e7c4119a/a3d5b93e-1b76-48e7-bb3e-2423e7c4119a.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 837.168217] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cef19620-c81c-4919-ad42-df7299df0b7c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.178597] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 837.178597] env[63372]: value = "task-1024237" [ 837.178597] env[63372]: _type = "Task" [ 837.178597] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.189978] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024237, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.400605] env[63372]: DEBUG nova.compute.manager [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 837.409519] env[63372]: DEBUG nova.network.neutron [-] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.441764] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 837.442952] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-67d8e706-a84a-4c38-8b58-f26c82c3b444 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.456527] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 837.456527] env[63372]: value = "task-1024238" [ 837.456527] env[63372]: _type = "Task" [ 837.456527] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.467284] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024238, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.510199] env[63372]: DEBUG nova.compute.manager [req-ac1002ed-4306-4579-978e-f4c648c912fc req-840191b8-7879-4142-84a3-c09c927fc260 service nova] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Received event network-vif-deleted-63d3f72b-cd36-4e27-bfac-e43665ed8ca1 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.516183] env[63372]: DEBUG nova.compute.manager [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 837.689584] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024237, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.747810] env[63372]: DEBUG nova.network.neutron [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Successfully created port: d27ee5ea-fd81-4901-bdf2-9876321872b5 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.809457] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "4028d7c0-f398-4e43-9cff-5d89a14c4efa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.809733] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "4028d7c0-f398-4e43-9cff-5d89a14c4efa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.809938] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "4028d7c0-f398-4e43-9cff-5d89a14c4efa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.810212] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "4028d7c0-f398-4e43-9cff-5d89a14c4efa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.811071] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "4028d7c0-f398-4e43-9cff-5d89a14c4efa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.817362] env[63372]: INFO nova.compute.manager [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Terminating instance [ 837.819687] env[63372]: DEBUG nova.compute.manager [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 837.819942] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 837.820849] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7110007f-9472-4bbc-8aad-29563625cf54 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.831659] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 837.831960] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72091488-2d50-42ca-90cc-992aa45bce54 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.836099] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31f2813-be78-462d-a420-3d2266a40f47 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.840422] env[63372]: DEBUG oslo_vmware.api [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 837.840422] env[63372]: value = "task-1024239" [ 837.840422] env[63372]: _type = "Task" [ 837.840422] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.847320] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f090d4f-c486-4452-9f3c-8d301279816d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.854581] env[63372]: DEBUG oslo_vmware.api [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024239, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.885867] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c9df9c-b23c-4a36-b652-681a3396d5ee {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.893163] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b530f29-bade-4c68-bfa7-2ddf9dbb83e6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.908896] env[63372]: DEBUG nova.compute.provider_tree [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.914695] env[63372]: INFO nova.compute.manager [-] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Took 1.41 seconds to deallocate network for instance. [ 837.967651] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024238, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.043589] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.189049] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024237, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.621398} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.191090] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] a3d5b93e-1b76-48e7-bb3e-2423e7c4119a/a3d5b93e-1b76-48e7-bb3e-2423e7c4119a.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 838.191090] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.191090] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-11e0f674-badc-4082-9755-fc6a44de8299 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.198968] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 838.198968] env[63372]: value = "task-1024240" [ 838.198968] env[63372]: _type = "Task" [ 838.198968] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.208572] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024240, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.352842] env[63372]: DEBUG oslo_vmware.api [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024239, 'name': PowerOffVM_Task, 'duration_secs': 0.233818} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.353121] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 838.353327] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 838.353585] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97eef17c-62cf-4641-ae7f-3c9977adbf7d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.412467] env[63372]: DEBUG nova.scheduler.client.report [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.417286] env[63372]: DEBUG nova.compute.manager [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 838.422652] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.430689] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 838.430940] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 838.431202] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleting the datastore file [datastore1] 4028d7c0-f398-4e43-9cff-5d89a14c4efa {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 838.434129] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-544611cc-f589-4d42-b368-a2671e31688f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.442018] env[63372]: DEBUG oslo_vmware.api [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 838.442018] env[63372]: value = "task-1024242" [ 838.442018] env[63372]: _type = "Task" [ 838.442018] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.452130] env[63372]: DEBUG oslo_vmware.api [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024242, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.454396] env[63372]: DEBUG nova.virt.hardware [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 838.454885] env[63372]: DEBUG nova.virt.hardware [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 838.455267] env[63372]: DEBUG nova.virt.hardware [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.455581] env[63372]: DEBUG nova.virt.hardware [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 838.455892] env[63372]: DEBUG nova.virt.hardware [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.456196] env[63372]: DEBUG nova.virt.hardware [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 838.456563] env[63372]: DEBUG nova.virt.hardware [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 838.456844] env[63372]: DEBUG nova.virt.hardware [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 838.457153] env[63372]: DEBUG nova.virt.hardware [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 838.457460] env[63372]: DEBUG nova.virt.hardware [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 838.461019] env[63372]: DEBUG nova.virt.hardware [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 838.461019] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-096b73b0-6bea-4fba-9436-1249723488be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.475678] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e8e3ed-8058-47bd-8a5c-f23ab10107e3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.480206] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024238, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.711836] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024240, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068156} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.712083] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.712916] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c75d04e-5ca1-44d7-ae65-a9bc0dbac4b3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.738396] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] a3d5b93e-1b76-48e7-bb3e-2423e7c4119a/a3d5b93e-1b76-48e7-bb3e-2423e7c4119a.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.738652] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48c22eda-048a-4f2d-8e1f-224af187b6d9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.760888] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 838.760888] env[63372]: value = "task-1024243" [ 838.760888] env[63372]: _type = "Task" [ 838.760888] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.769469] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024243, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.921966] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.539s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.922537] env[63372]: DEBUG nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 838.925187] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.718s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.926563] env[63372]: INFO nova.compute.claims [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.952071] env[63372]: DEBUG oslo_vmware.api [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024242, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.972650] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024238, 'name': CloneVM_Task} progress is 95%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.274844] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024243, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.437062] env[63372]: DEBUG nova.compute.utils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.438793] env[63372]: DEBUG nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 839.438967] env[63372]: DEBUG nova.network.neutron [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 839.461456] env[63372]: DEBUG oslo_vmware.api [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024242, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.589892} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.469442] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 839.469794] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 839.469979] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 839.470178] env[63372]: INFO nova.compute.manager [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Took 1.65 seconds to destroy the instance on the hypervisor. [ 839.470638] env[63372]: DEBUG oslo.service.loopingcall [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 839.471454] env[63372]: DEBUG nova.compute.manager [-] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 839.471632] env[63372]: DEBUG nova.network.neutron [-] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 839.479794] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024238, 'name': CloneVM_Task, 'duration_secs': 1.759114} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.480067] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Created linked-clone VM from snapshot [ 839.480850] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06149fc2-f9f1-4e26-90f0-49d85d2562a9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.489932] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Uploading image 217465cd-6c37-4aef-bacb-fbc93077f64b {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 839.507214] env[63372]: DEBUG nova.policy [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3d8052fc4294245bec22af4c254fce0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41ca5f8268434d329deec67fb2867b59', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 839.521064] env[63372]: DEBUG oslo_vmware.rw_handles [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 839.521064] env[63372]: value = "vm-227376" [ 839.521064] env[63372]: _type = "VirtualMachine" [ 839.521064] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 839.521402] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-6178d859-561a-4d8e-a3e1-04dfb5c382be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.530881] env[63372]: DEBUG oslo_vmware.rw_handles [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lease: (returnval){ [ 839.530881] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b582b5-3e48-129b-9b5f-380cd3bdc5d9" [ 839.530881] env[63372]: _type = "HttpNfcLease" [ 839.530881] env[63372]: } obtained for exporting VM: (result){ [ 839.530881] env[63372]: value = "vm-227376" [ 839.530881] env[63372]: _type = "VirtualMachine" [ 839.530881] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 839.531259] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the lease: (returnval){ [ 839.531259] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b582b5-3e48-129b-9b5f-380cd3bdc5d9" [ 839.531259] env[63372]: _type = "HttpNfcLease" [ 839.531259] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 839.543763] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 839.543763] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b582b5-3e48-129b-9b5f-380cd3bdc5d9" [ 839.543763] env[63372]: _type = "HttpNfcLease" [ 839.543763] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 839.607117] env[63372]: DEBUG nova.compute.manager [req-74eecd5e-0519-4075-9257-87803285e42a req-820cdf25-c281-4c99-b4a7-daf256b2cf1d service nova] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Received event network-vif-plugged-d27ee5ea-fd81-4901-bdf2-9876321872b5 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 839.607117] env[63372]: DEBUG oslo_concurrency.lockutils [req-74eecd5e-0519-4075-9257-87803285e42a req-820cdf25-c281-4c99-b4a7-daf256b2cf1d service nova] Acquiring lock "07dab964-a741-4723-98a3-0cfbe4eff975-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.609194] env[63372]: DEBUG oslo_concurrency.lockutils [req-74eecd5e-0519-4075-9257-87803285e42a req-820cdf25-c281-4c99-b4a7-daf256b2cf1d service nova] Lock "07dab964-a741-4723-98a3-0cfbe4eff975-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.609194] env[63372]: DEBUG oslo_concurrency.lockutils [req-74eecd5e-0519-4075-9257-87803285e42a req-820cdf25-c281-4c99-b4a7-daf256b2cf1d service nova] Lock "07dab964-a741-4723-98a3-0cfbe4eff975-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.609194] env[63372]: DEBUG nova.compute.manager [req-74eecd5e-0519-4075-9257-87803285e42a req-820cdf25-c281-4c99-b4a7-daf256b2cf1d service nova] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] No waiting events found dispatching network-vif-plugged-d27ee5ea-fd81-4901-bdf2-9876321872b5 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 839.609194] env[63372]: WARNING nova.compute.manager [req-74eecd5e-0519-4075-9257-87803285e42a req-820cdf25-c281-4c99-b4a7-daf256b2cf1d service nova] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Received unexpected event network-vif-plugged-d27ee5ea-fd81-4901-bdf2-9876321872b5 for instance with vm_state building and task_state spawning. [ 839.777093] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024243, 'name': ReconfigVM_Task, 'duration_secs': 0.791648} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.777513] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Reconfigured VM instance instance-00000046 to attach disk [datastore2] a3d5b93e-1b76-48e7-bb3e-2423e7c4119a/a3d5b93e-1b76-48e7-bb3e-2423e7c4119a.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.778249] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2173cb19-3f76-4e3d-b3ec-b414a36b9b5f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.786446] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 839.786446] env[63372]: value = "task-1024245" [ 839.786446] env[63372]: _type = "Task" [ 839.786446] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.797402] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024245, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.804050] env[63372]: DEBUG nova.network.neutron [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Successfully created port: bfc2f0a4-c40a-49e2-9d87-e7f9593a583a {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 839.946769] env[63372]: DEBUG nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 840.043152] env[63372]: DEBUG nova.network.neutron [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Successfully updated port: d27ee5ea-fd81-4901-bdf2-9876321872b5 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.050020] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 840.050020] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b582b5-3e48-129b-9b5f-380cd3bdc5d9" [ 840.050020] env[63372]: _type = "HttpNfcLease" [ 840.050020] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 840.050809] env[63372]: DEBUG oslo_vmware.rw_handles [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 840.050809] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b582b5-3e48-129b-9b5f-380cd3bdc5d9" [ 840.050809] env[63372]: _type = "HttpNfcLease" [ 840.050809] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 840.051297] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a253d2-0275-49eb-b56d-4a110889ac65 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.069641] env[63372]: DEBUG oslo_vmware.rw_handles [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f900f0-cd4e-45e0-f03b-6146fd73591c/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 840.069851] env[63372]: DEBUG oslo_vmware.rw_handles [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f900f0-cd4e-45e0-f03b-6146fd73591c/disk-0.vmdk for reading. {{(pid=63372) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 840.170595] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-11b7c480-d04e-464b-b295-829c75147586 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.298255] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024245, 'name': Rename_Task, 'duration_secs': 0.159565} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.299055] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 840.299055] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-08728b97-0cdf-45e2-a540-11acfed9b32f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.307292] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 840.307292] env[63372]: value = "task-1024246" [ 840.307292] env[63372]: _type = "Task" [ 840.307292] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.318577] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024246, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.444746] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-537b7bcd-a9b1-4f8f-ab6e-53813093a0ac {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.464488] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c4fe44-c35f-4d15-896f-36a4895f4ae3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.502614] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc0cacea-1b1d-4b5c-b0b3-bc7a58a5753f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.513601] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee832ebc-06fc-463c-a333-7eb0eb3e7e9b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.531199] env[63372]: DEBUG nova.compute.provider_tree [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.546225] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Acquiring lock "refresh_cache-07dab964-a741-4723-98a3-0cfbe4eff975" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.546225] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Acquired lock "refresh_cache-07dab964-a741-4723-98a3-0cfbe4eff975" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.546389] env[63372]: DEBUG nova.network.neutron [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 840.641221] env[63372]: DEBUG nova.network.neutron [-] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.821885] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024246, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.958496] env[63372]: DEBUG nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 840.984733] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 840.985163] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 840.985408] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.985868] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 840.986121] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.986324] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 840.988803] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 840.988803] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 840.988803] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 840.988803] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 840.988803] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 840.988803] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f719841a-a4fd-4290-a822-2fbfe0d7eeb8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.999190] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24dfb817-f430-42f3-aad1-4fbb079e7250 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.035640] env[63372]: DEBUG nova.scheduler.client.report [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 841.084105] env[63372]: DEBUG nova.network.neutron [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 841.145029] env[63372]: INFO nova.compute.manager [-] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Took 1.67 seconds to deallocate network for instance. [ 841.321573] env[63372]: DEBUG nova.network.neutron [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Updating instance_info_cache with network_info: [{"id": "d27ee5ea-fd81-4901-bdf2-9876321872b5", "address": "fa:16:3e:f3:aa:69", "network": {"id": "76c403d1-8445-4a29-9b40-69e7ab69d7f6", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-980316045-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "56e781c01b95446b91c6beae143d1be4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd27ee5ea-fd", "ovs_interfaceid": "d27ee5ea-fd81-4901-bdf2-9876321872b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.329705] env[63372]: DEBUG oslo_vmware.api [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024246, 'name': PowerOnVM_Task, 'duration_secs': 0.726386} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.330186] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 841.330583] env[63372]: INFO nova.compute.manager [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Took 10.65 seconds to spawn the instance on the hypervisor. [ 841.330769] env[63372]: DEBUG nova.compute.manager [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.331724] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcf00af3-2f32-453c-a035-6e02c92e5915 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.492888] env[63372]: DEBUG nova.network.neutron [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Successfully updated port: bfc2f0a4-c40a-49e2-9d87-e7f9593a583a {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 841.540750] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.615s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.541383] env[63372]: DEBUG nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 841.544264] env[63372]: DEBUG oslo_concurrency.lockutils [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.416s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.544500] env[63372]: DEBUG nova.objects.instance [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lazy-loading 'resources' on Instance uuid 7c9d551e-d210-4943-971f-b2829751fcfc {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.655077] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.743339] env[63372]: DEBUG nova.compute.manager [req-a3c12648-328e-47e1-a093-e0a4542070a3 req-1abdda28-5942-4449-b6a6-57e8a3f36bd4 service nova] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Received event network-vif-plugged-bfc2f0a4-c40a-49e2-9d87-e7f9593a583a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.743339] env[63372]: DEBUG oslo_concurrency.lockutils [req-a3c12648-328e-47e1-a093-e0a4542070a3 req-1abdda28-5942-4449-b6a6-57e8a3f36bd4 service nova] Acquiring lock "798047e4-1a2b-458c-86c2-5567197d1861-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.743339] env[63372]: DEBUG oslo_concurrency.lockutils [req-a3c12648-328e-47e1-a093-e0a4542070a3 req-1abdda28-5942-4449-b6a6-57e8a3f36bd4 service nova] Lock "798047e4-1a2b-458c-86c2-5567197d1861-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.743339] env[63372]: DEBUG oslo_concurrency.lockutils [req-a3c12648-328e-47e1-a093-e0a4542070a3 req-1abdda28-5942-4449-b6a6-57e8a3f36bd4 service nova] Lock "798047e4-1a2b-458c-86c2-5567197d1861-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.743339] env[63372]: DEBUG nova.compute.manager [req-a3c12648-328e-47e1-a093-e0a4542070a3 req-1abdda28-5942-4449-b6a6-57e8a3f36bd4 service nova] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] No waiting events found dispatching network-vif-plugged-bfc2f0a4-c40a-49e2-9d87-e7f9593a583a {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 841.743339] env[63372]: WARNING nova.compute.manager [req-a3c12648-328e-47e1-a093-e0a4542070a3 req-1abdda28-5942-4449-b6a6-57e8a3f36bd4 service nova] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Received unexpected event network-vif-plugged-bfc2f0a4-c40a-49e2-9d87-e7f9593a583a for instance with vm_state building and task_state spawning. [ 841.825077] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Releasing lock "refresh_cache-07dab964-a741-4723-98a3-0cfbe4eff975" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.825934] env[63372]: DEBUG nova.compute.manager [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Instance network_info: |[{"id": "d27ee5ea-fd81-4901-bdf2-9876321872b5", "address": "fa:16:3e:f3:aa:69", "network": {"id": "76c403d1-8445-4a29-9b40-69e7ab69d7f6", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-980316045-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "56e781c01b95446b91c6beae143d1be4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd27ee5ea-fd", "ovs_interfaceid": "d27ee5ea-fd81-4901-bdf2-9876321872b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 841.826647] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:aa:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '456bd8a2-0fb6-4b17-9d25-08e7995c5184', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd27ee5ea-fd81-4901-bdf2-9876321872b5', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.835508] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Creating folder: Project (56e781c01b95446b91c6beae143d1be4). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 841.836029] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-229d256c-835c-41a3-ac25-d6f80c1bb539 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.853104] env[63372]: INFO nova.compute.manager [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Took 46.51 seconds to build instance. [ 841.856981] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Created folder: Project (56e781c01b95446b91c6beae143d1be4) in parent group-v227230. [ 841.858325] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Creating folder: Instances. Parent ref: group-v227377. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 841.859806] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9b34fec4-f84b-4b59-bc7a-7ca1c14463de {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.876875] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Created folder: Instances in parent group-v227377. [ 841.878296] env[63372]: DEBUG oslo.service.loopingcall [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.878296] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 841.878296] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f70dc580-c7fc-4398-9ac2-b16a1c2ddc92 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.915516] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.915516] env[63372]: value = "task-1024249" [ 841.915516] env[63372]: _type = "Task" [ 841.915516] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.928746] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024249, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.993174] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "refresh_cache-798047e4-1a2b-458c-86c2-5567197d1861" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.993174] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired lock "refresh_cache-798047e4-1a2b-458c-86c2-5567197d1861" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.993174] env[63372]: DEBUG nova.network.neutron [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 842.027426] env[63372]: DEBUG nova.compute.manager [req-a406c7d3-d29b-43ed-bc52-adb062b820a1 req-d3c383cf-65b7-483a-a5d3-ddd255f1e19d service nova] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Received event network-changed-d27ee5ea-fd81-4901-bdf2-9876321872b5 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.027709] env[63372]: DEBUG nova.compute.manager [req-a406c7d3-d29b-43ed-bc52-adb062b820a1 req-d3c383cf-65b7-483a-a5d3-ddd255f1e19d service nova] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Refreshing instance network info cache due to event network-changed-d27ee5ea-fd81-4901-bdf2-9876321872b5. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 842.028110] env[63372]: DEBUG oslo_concurrency.lockutils [req-a406c7d3-d29b-43ed-bc52-adb062b820a1 req-d3c383cf-65b7-483a-a5d3-ddd255f1e19d service nova] Acquiring lock "refresh_cache-07dab964-a741-4723-98a3-0cfbe4eff975" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.028411] env[63372]: DEBUG oslo_concurrency.lockutils [req-a406c7d3-d29b-43ed-bc52-adb062b820a1 req-d3c383cf-65b7-483a-a5d3-ddd255f1e19d service nova] Acquired lock "refresh_cache-07dab964-a741-4723-98a3-0cfbe4eff975" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.028696] env[63372]: DEBUG nova.network.neutron [req-a406c7d3-d29b-43ed-bc52-adb062b820a1 req-d3c383cf-65b7-483a-a5d3-ddd255f1e19d service nova] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Refreshing network info cache for port d27ee5ea-fd81-4901-bdf2-9876321872b5 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 842.050584] env[63372]: DEBUG nova.compute.utils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 842.060326] env[63372]: DEBUG nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 842.060549] env[63372]: DEBUG nova.network.neutron [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 842.084368] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "c01a5d24-eb46-4a69-993e-753880ce8e85" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.084926] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.084926] env[63372]: INFO nova.compute.manager [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Shelving [ 842.115037] env[63372]: DEBUG nova.policy [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3d8052fc4294245bec22af4c254fce0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41ca5f8268434d329deec67fb2867b59', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 842.158711] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquiring lock "1f52f9ee-2af3-4321-8ec6-5d5e841ed880" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.158711] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Lock "1f52f9ee-2af3-4321-8ec6-5d5e841ed880" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.360141] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1880372-f081-4cad-8728-a22ecda5116e tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "a3d5b93e-1b76-48e7-bb3e-2423e7c4119a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.025s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.430203] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024249, 'name': CreateVM_Task, 'duration_secs': 0.448287} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.430400] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 842.431175] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.431342] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.431664] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 842.432040] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8063d9c3-f09c-403e-ad29-32c15e4b955b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.441175] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Waiting for the task: (returnval){ [ 842.441175] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520db81e-21fd-7f33-f608-8bff65511b72" [ 842.441175] env[63372]: _type = "Task" [ 842.441175] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.452918] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520db81e-21fd-7f33-f608-8bff65511b72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.521279] env[63372]: DEBUG nova.network.neutron [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Successfully created port: a80c4949-9726-4644-967c-f72b56d42665 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 842.522943] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f830fb5-3d3e-49b6-84ac-0220cd472e1b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.534339] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9170bfaf-ecf6-447e-845c-5cec89307ad6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.570214] env[63372]: DEBUG nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 842.574097] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44bfff8-a93e-440b-8bad-aeaad20bfde1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.578943] env[63372]: DEBUG nova.network.neutron [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 842.587199] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6c41890-41cf-4f34-903f-701a9a35fc6b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.607748] env[63372]: DEBUG nova.compute.provider_tree [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.611420] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 842.614595] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2e1caa9-6994-4619-ad6d-86b751aeb922 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.621145] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 842.621145] env[63372]: value = "task-1024250" [ 842.621145] env[63372]: _type = "Task" [ 842.621145] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.635533] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024250, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.663696] env[63372]: DEBUG nova.compute.manager [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 842.956085] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520db81e-21fd-7f33-f608-8bff65511b72, 'name': SearchDatastore_Task, 'duration_secs': 0.01375} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.957148] env[63372]: DEBUG nova.network.neutron [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Updating instance_info_cache with network_info: [{"id": "bfc2f0a4-c40a-49e2-9d87-e7f9593a583a", "address": "fa:16:3e:5a:b5:23", "network": {"id": "49776280-f4e5-4f22-98fd-a6e6ff70c80f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1268791360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41ca5f8268434d329deec67fb2867b59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfc2f0a4-c4", "ovs_interfaceid": "bfc2f0a4-c40a-49e2-9d87-e7f9593a583a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.961938] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.961938] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 842.961938] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.961938] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.961938] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 842.961938] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd5c8319-a5a0-4b46-96fe-c1f6d71acbf8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.971749] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 842.971990] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 842.972717] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cb3cd85-5f05-4261-add6-515ad0e06c51 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.979364] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Waiting for the task: (returnval){ [ 842.979364] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52c09515-0c58-3c3e-1562-5b6d0d7df872" [ 842.979364] env[63372]: _type = "Task" [ 842.979364] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.989180] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c09515-0c58-3c3e-1562-5b6d0d7df872, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.028579] env[63372]: DEBUG nova.network.neutron [req-a406c7d3-d29b-43ed-bc52-adb062b820a1 req-d3c383cf-65b7-483a-a5d3-ddd255f1e19d service nova] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Updated VIF entry in instance network info cache for port d27ee5ea-fd81-4901-bdf2-9876321872b5. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 843.028960] env[63372]: DEBUG nova.network.neutron [req-a406c7d3-d29b-43ed-bc52-adb062b820a1 req-d3c383cf-65b7-483a-a5d3-ddd255f1e19d service nova] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Updating instance_info_cache with network_info: [{"id": "d27ee5ea-fd81-4901-bdf2-9876321872b5", "address": "fa:16:3e:f3:aa:69", "network": {"id": "76c403d1-8445-4a29-9b40-69e7ab69d7f6", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-980316045-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "56e781c01b95446b91c6beae143d1be4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd27ee5ea-fd", "ovs_interfaceid": "d27ee5ea-fd81-4901-bdf2-9876321872b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.115297] env[63372]: DEBUG nova.scheduler.client.report [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 843.132996] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024250, 'name': PowerOffVM_Task, 'duration_secs': 0.319914} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.133195] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 843.134015] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e37a68-6372-4949-b177-15dd3911dd00 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.155250] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795fcf5d-347b-4490-aa8a-9976e5ff05ce {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.186214] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.461309] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Releasing lock "refresh_cache-798047e4-1a2b-458c-86c2-5567197d1861" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.461645] env[63372]: DEBUG nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Instance network_info: |[{"id": "bfc2f0a4-c40a-49e2-9d87-e7f9593a583a", "address": "fa:16:3e:5a:b5:23", "network": {"id": "49776280-f4e5-4f22-98fd-a6e6ff70c80f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1268791360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41ca5f8268434d329deec67fb2867b59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfc2f0a4-c4", "ovs_interfaceid": "bfc2f0a4-c40a-49e2-9d87-e7f9593a583a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 843.462091] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:b5:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffcecdaa-a7b8-49fc-9371-dbdb7744688e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bfc2f0a4-c40a-49e2-9d87-e7f9593a583a', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.470351] env[63372]: DEBUG oslo.service.loopingcall [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.470946] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 843.471212] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d61a793f-9b59-4887-af75-e22d1e622355 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.496344] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c09515-0c58-3c3e-1562-5b6d0d7df872, 'name': SearchDatastore_Task, 'duration_secs': 0.016557} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.498096] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.498096] env[63372]: value = "task-1024251" [ 843.498096] env[63372]: _type = "Task" [ 843.498096] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.498325] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39756a0b-a1c7-412c-9b5b-77b4d1ce18d8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.511036] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024251, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.512084] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Waiting for the task: (returnval){ [ 843.512084] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520e231b-646e-50ff-60ab-10c0111d94e4" [ 843.512084] env[63372]: _type = "Task" [ 843.512084] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.520870] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520e231b-646e-50ff-60ab-10c0111d94e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.531644] env[63372]: DEBUG oslo_concurrency.lockutils [req-a406c7d3-d29b-43ed-bc52-adb062b820a1 req-d3c383cf-65b7-483a-a5d3-ddd255f1e19d service nova] Releasing lock "refresh_cache-07dab964-a741-4723-98a3-0cfbe4eff975" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.531928] env[63372]: DEBUG nova.compute.manager [req-a406c7d3-d29b-43ed-bc52-adb062b820a1 req-d3c383cf-65b7-483a-a5d3-ddd255f1e19d service nova] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Received event network-vif-deleted-95bb06d1-3333-43b4-a667-f99370acc1d7 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.586261] env[63372]: DEBUG nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 843.626758] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.627033] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.627247] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.627451] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.627609] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.627814] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.628050] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.628283] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.628463] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.628645] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.628872] env[63372]: DEBUG nova.virt.hardware [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.629721] env[63372]: DEBUG oslo_concurrency.lockutils [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.085s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.632567] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee98eff-e3c1-436c-b909-e829b4fe9d00 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.635491] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.708s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.635732] env[63372]: DEBUG nova.objects.instance [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lazy-loading 'resources' on Instance uuid 368101ee-e4fd-4fe4-b7c8-75c6b210189f {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.644629] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048a80fb-6e2c-40cd-8eb9-78355641c705 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.662771] env[63372]: INFO nova.scheduler.client.report [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Deleted allocations for instance 7c9d551e-d210-4943-971f-b2829751fcfc [ 843.667740] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 843.668133] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e1a67979-4741-464b-9aa8-9d5f496be7b3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.678576] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 843.678576] env[63372]: value = "task-1024252" [ 843.678576] env[63372]: _type = "Task" [ 843.678576] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.693448] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024252, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.774229] env[63372]: DEBUG nova.compute.manager [req-1146a50a-138a-4633-9e32-3f9668212a43 req-c44ac3ec-9e62-44bc-b9ad-9a19dc9066b1 service nova] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Received event network-changed-bfc2f0a4-c40a-49e2-9d87-e7f9593a583a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.774442] env[63372]: DEBUG nova.compute.manager [req-1146a50a-138a-4633-9e32-3f9668212a43 req-c44ac3ec-9e62-44bc-b9ad-9a19dc9066b1 service nova] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Refreshing instance network info cache due to event network-changed-bfc2f0a4-c40a-49e2-9d87-e7f9593a583a. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 843.774652] env[63372]: DEBUG oslo_concurrency.lockutils [req-1146a50a-138a-4633-9e32-3f9668212a43 req-c44ac3ec-9e62-44bc-b9ad-9a19dc9066b1 service nova] Acquiring lock "refresh_cache-798047e4-1a2b-458c-86c2-5567197d1861" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.774808] env[63372]: DEBUG oslo_concurrency.lockutils [req-1146a50a-138a-4633-9e32-3f9668212a43 req-c44ac3ec-9e62-44bc-b9ad-9a19dc9066b1 service nova] Acquired lock "refresh_cache-798047e4-1a2b-458c-86c2-5567197d1861" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.774994] env[63372]: DEBUG nova.network.neutron [req-1146a50a-138a-4633-9e32-3f9668212a43 req-c44ac3ec-9e62-44bc-b9ad-9a19dc9066b1 service nova] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Refreshing network info cache for port bfc2f0a4-c40a-49e2-9d87-e7f9593a583a {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 844.014660] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024251, 'name': CreateVM_Task, 'duration_secs': 0.468691} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.018079] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 844.019381] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.019381] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.019948] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.020253] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9dd572e5-9825-406a-9fda-4e07c912ea82 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.027656] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520e231b-646e-50ff-60ab-10c0111d94e4, 'name': SearchDatastore_Task, 'duration_secs': 0.01705} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.028987] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.029127] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 07dab964-a741-4723-98a3-0cfbe4eff975/07dab964-a741-4723-98a3-0cfbe4eff975.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 844.029517] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 844.029517] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5275a894-8c11-1263-530b-39b7bf00f0d5" [ 844.029517] env[63372]: _type = "Task" [ 844.029517] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.029744] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc8c72c0-f003-47d7-a68b-ec76503d35b6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.042302] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5275a894-8c11-1263-530b-39b7bf00f0d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.044754] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Waiting for the task: (returnval){ [ 844.044754] env[63372]: value = "task-1024253" [ 844.044754] env[63372]: _type = "Task" [ 844.044754] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.056582] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024253, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.143342] env[63372]: DEBUG nova.network.neutron [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Successfully updated port: a80c4949-9726-4644-967c-f72b56d42665 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 844.173878] env[63372]: DEBUG oslo_concurrency.lockutils [None req-27fdf2ac-557a-47c2-a4e0-7c2b1be4d9dd tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "7c9d551e-d210-4943-971f-b2829751fcfc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.700s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.196213] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024252, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.520823] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1c0679-62d0-4a3a-9927-351cd18582ce {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.530446] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650c9558-72e2-429b-af50-5937d376d995 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.546355] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5275a894-8c11-1263-530b-39b7bf00f0d5, 'name': SearchDatastore_Task, 'duration_secs': 0.017695} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.575394] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.575677] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.575916] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.576155] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.576321] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.579941] env[63372]: DEBUG nova.network.neutron [req-1146a50a-138a-4633-9e32-3f9668212a43 req-c44ac3ec-9e62-44bc-b9ad-9a19dc9066b1 service nova] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Updated VIF entry in instance network info cache for port bfc2f0a4-c40a-49e2-9d87-e7f9593a583a. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 844.580180] env[63372]: DEBUG nova.network.neutron [req-1146a50a-138a-4633-9e32-3f9668212a43 req-c44ac3ec-9e62-44bc-b9ad-9a19dc9066b1 service nova] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Updating instance_info_cache with network_info: [{"id": "bfc2f0a4-c40a-49e2-9d87-e7f9593a583a", "address": "fa:16:3e:5a:b5:23", "network": {"id": "49776280-f4e5-4f22-98fd-a6e6ff70c80f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1268791360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41ca5f8268434d329deec67fb2867b59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfc2f0a4-c4", "ovs_interfaceid": "bfc2f0a4-c40a-49e2-9d87-e7f9593a583a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.581434] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-acfbe772-b111-4e8c-a660-20e6e7ea26aa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.584114] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98deac41-ba32-4a50-8b45-aec98b60a18d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.588947] env[63372]: DEBUG oslo_concurrency.lockutils [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "de1b38a4-c7f2-420c-a050-7311976e4ca8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.589167] env[63372]: DEBUG oslo_concurrency.lockutils [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "de1b38a4-c7f2-420c-a050-7311976e4ca8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.589428] env[63372]: DEBUG oslo_concurrency.lockutils [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "de1b38a4-c7f2-420c-a050-7311976e4ca8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.589641] env[63372]: DEBUG oslo_concurrency.lockutils [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "de1b38a4-c7f2-420c-a050-7311976e4ca8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.590088] env[63372]: DEBUG oslo_concurrency.lockutils [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "de1b38a4-c7f2-420c-a050-7311976e4ca8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.597933] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024253, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.599056] env[63372]: INFO nova.compute.manager [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Terminating instance [ 844.601895] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8cd9eb-a505-4b11-b7c1-68074cf3a53f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.607922] env[63372]: DEBUG nova.compute.manager [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 844.607922] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 844.609018] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa0fe08-2642-44d1-b1e5-9d89a25bdbfe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.613911] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.614133] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 844.615499] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7f55f6b-0def-4666-be0d-2cba5dd48042 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.630216] env[63372]: DEBUG nova.compute.provider_tree [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.631486] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 844.632406] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-57d86225-b9d7-425c-b651-5efd2df72fbc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.636627] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 844.636627] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fa7d87-a59d-85ff-cc66-e3b1835927e1" [ 844.636627] env[63372]: _type = "Task" [ 844.636627] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.642943] env[63372]: DEBUG oslo_vmware.api [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 844.642943] env[63372]: value = "task-1024254" [ 844.642943] env[63372]: _type = "Task" [ 844.642943] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.646856] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fa7d87-a59d-85ff-cc66-e3b1835927e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.650688] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "refresh_cache-1d047728-50d8-465c-b217-6cbe9bbea3e8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.650825] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired lock "refresh_cache-1d047728-50d8-465c-b217-6cbe9bbea3e8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.650971] env[63372]: DEBUG nova.network.neutron [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 844.658507] env[63372]: DEBUG oslo_vmware.api [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024254, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.692110] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024252, 'name': CreateSnapshot_Task, 'duration_secs': 0.89219} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.692411] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 844.693496] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762572da-05f7-4adc-881f-06f539ed91bf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.057523] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024253, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.685805} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.057872] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 07dab964-a741-4723-98a3-0cfbe4eff975/07dab964-a741-4723-98a3-0cfbe4eff975.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 845.058193] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.058516] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df344d61-7544-4f84-9d33-8545e087aca3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.066967] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Waiting for the task: (returnval){ [ 845.066967] env[63372]: value = "task-1024255" [ 845.066967] env[63372]: _type = "Task" [ 845.066967] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.076988] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024255, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.087829] env[63372]: DEBUG oslo_concurrency.lockutils [req-1146a50a-138a-4633-9e32-3f9668212a43 req-c44ac3ec-9e62-44bc-b9ad-9a19dc9066b1 service nova] Releasing lock "refresh_cache-798047e4-1a2b-458c-86c2-5567197d1861" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.088473] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "a3d5b93e-1b76-48e7-bb3e-2423e7c4119a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.088701] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "a3d5b93e-1b76-48e7-bb3e-2423e7c4119a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.089013] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "a3d5b93e-1b76-48e7-bb3e-2423e7c4119a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.089426] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "a3d5b93e-1b76-48e7-bb3e-2423e7c4119a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.089655] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "a3d5b93e-1b76-48e7-bb3e-2423e7c4119a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.092126] env[63372]: INFO nova.compute.manager [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Terminating instance [ 845.094338] env[63372]: DEBUG nova.compute.manager [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 845.094602] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 845.095475] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d24964-9cee-4eed-b660-b1ed3d2739d2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.103816] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 845.104099] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1a591a8-0b38-41d4-915d-3f28d0630f33 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.111266] env[63372]: DEBUG oslo_vmware.api [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 845.111266] env[63372]: value = "task-1024256" [ 845.111266] env[63372]: _type = "Task" [ 845.111266] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.120822] env[63372]: DEBUG oslo_vmware.api [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024256, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.133543] env[63372]: DEBUG nova.scheduler.client.report [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.147436] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fa7d87-a59d-85ff-cc66-e3b1835927e1, 'name': SearchDatastore_Task, 'duration_secs': 0.051286} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.151229] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df931482-6e3f-497a-953a-aae3c1ed4895 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.161028] env[63372]: DEBUG oslo_vmware.api [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024254, 'name': PowerOffVM_Task, 'duration_secs': 0.403534} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.162363] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 845.162625] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 845.162914] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 845.162914] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5218aff3-cc54-fc80-364e-5bdb626ff928" [ 845.162914] env[63372]: _type = "Task" [ 845.162914] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.163187] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c3ae3063-cd44-4fa2-8eb1-2d6286ca1ed5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.174853] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5218aff3-cc54-fc80-364e-5bdb626ff928, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.197032] env[63372]: DEBUG nova.network.neutron [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 845.212496] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 845.212833] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1402c89d-783b-4777-9aee-ff6a8ef4def0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.224300] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 845.224300] env[63372]: value = "task-1024258" [ 845.224300] env[63372]: _type = "Task" [ 845.224300] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.234952] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024258, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.254563] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 845.254753] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 845.254809] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Deleting the datastore file [datastore1] de1b38a4-c7f2-420c-a050-7311976e4ca8 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 845.255238] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ceaf9707-2c69-4ca2-9327-a2a03c475b8f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.265425] env[63372]: DEBUG oslo_vmware.api [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 845.265425] env[63372]: value = "task-1024259" [ 845.265425] env[63372]: _type = "Task" [ 845.265425] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.275868] env[63372]: DEBUG oslo_vmware.api [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024259, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.364490] env[63372]: DEBUG nova.network.neutron [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Updating instance_info_cache with network_info: [{"id": "a80c4949-9726-4644-967c-f72b56d42665", "address": "fa:16:3e:c5:a2:4d", "network": {"id": "49776280-f4e5-4f22-98fd-a6e6ff70c80f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1268791360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41ca5f8268434d329deec67fb2867b59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa80c4949-97", "ovs_interfaceid": "a80c4949-9726-4644-967c-f72b56d42665", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.578605] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024255, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.14602} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.578997] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 845.579899] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf7251c-30e6-4fe5-aeb0-dd3526f3c519 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.606122] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 07dab964-a741-4723-98a3-0cfbe4eff975/07dab964-a741-4723-98a3-0cfbe4eff975.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.606622] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20380747-d9ef-430b-a61e-f7b2d312a653 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.632778] env[63372]: DEBUG oslo_vmware.api [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024256, 'name': PowerOffVM_Task, 'duration_secs': 0.474815} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.634447] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 845.634565] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 845.634942] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Waiting for the task: (returnval){ [ 845.634942] env[63372]: value = "task-1024260" [ 845.634942] env[63372]: _type = "Task" [ 845.634942] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.635272] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f3e604f-d7a9-4ddd-a66e-a7ac2cae6364 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.640673] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.005s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.643704] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.422s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.643854] env[63372]: DEBUG nova.objects.instance [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lazy-loading 'resources' on Instance uuid 8283b736-ad02-4082-97b7-561bd5c5da93 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.652451] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024260, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.674202] env[63372]: INFO nova.scheduler.client.report [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Deleted allocations for instance 368101ee-e4fd-4fe4-b7c8-75c6b210189f [ 845.681708] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5218aff3-cc54-fc80-364e-5bdb626ff928, 'name': SearchDatastore_Task, 'duration_secs': 0.0246} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.689293] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.689808] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 798047e4-1a2b-458c-86c2-5567197d1861/798047e4-1a2b-458c-86c2-5567197d1861.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 845.690942] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eab1440d-cd85-4c43-82e1-2087cd066e3a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.700806] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 845.700806] env[63372]: value = "task-1024262" [ 845.700806] env[63372]: _type = "Task" [ 845.700806] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.710320] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024262, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.731748] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 845.732009] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 845.732239] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Deleting the datastore file [datastore2] a3d5b93e-1b76-48e7-bb3e-2423e7c4119a {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 845.735806] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0fb868d0-2e5d-42c2-abfb-a5a3e6fb53bf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.738334] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024258, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.743863] env[63372]: DEBUG oslo_vmware.api [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 845.743863] env[63372]: value = "task-1024263" [ 845.743863] env[63372]: _type = "Task" [ 845.743863] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.753843] env[63372]: DEBUG oslo_vmware.api [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024263, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.776057] env[63372]: DEBUG oslo_vmware.api [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024259, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.283956} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.776365] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 845.776607] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 845.776873] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 845.777090] env[63372]: INFO nova.compute.manager [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Took 1.17 seconds to destroy the instance on the hypervisor. [ 845.780229] env[63372]: DEBUG oslo.service.loopingcall [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.780229] env[63372]: DEBUG nova.compute.manager [-] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 845.780229] env[63372]: DEBUG nova.network.neutron [-] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 845.867657] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Releasing lock "refresh_cache-1d047728-50d8-465c-b217-6cbe9bbea3e8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.867994] env[63372]: DEBUG nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Instance network_info: |[{"id": "a80c4949-9726-4644-967c-f72b56d42665", "address": "fa:16:3e:c5:a2:4d", "network": {"id": "49776280-f4e5-4f22-98fd-a6e6ff70c80f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1268791360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41ca5f8268434d329deec67fb2867b59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa80c4949-97", "ovs_interfaceid": "a80c4949-9726-4644-967c-f72b56d42665", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 845.868491] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:a2:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffcecdaa-a7b8-49fc-9371-dbdb7744688e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a80c4949-9726-4644-967c-f72b56d42665', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 845.878066] env[63372]: DEBUG oslo.service.loopingcall [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.878365] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 845.878599] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f6ba100-0247-4048-8478-a1c3973171f3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.901922] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.901922] env[63372]: value = "task-1024264" [ 845.901922] env[63372]: _type = "Task" [ 845.901922] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.911439] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024264, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.012281] env[63372]: DEBUG nova.compute.manager [req-ebf8f5ce-d221-4eeb-ae76-2101a2cd6e5f req-385de9c2-3321-41a9-847e-1a7a9197a449 service nova] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Received event network-vif-plugged-a80c4949-9726-4644-967c-f72b56d42665 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.012696] env[63372]: DEBUG oslo_concurrency.lockutils [req-ebf8f5ce-d221-4eeb-ae76-2101a2cd6e5f req-385de9c2-3321-41a9-847e-1a7a9197a449 service nova] Acquiring lock "1d047728-50d8-465c-b217-6cbe9bbea3e8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.013523] env[63372]: DEBUG oslo_concurrency.lockutils [req-ebf8f5ce-d221-4eeb-ae76-2101a2cd6e5f req-385de9c2-3321-41a9-847e-1a7a9197a449 service nova] Lock "1d047728-50d8-465c-b217-6cbe9bbea3e8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.013523] env[63372]: DEBUG oslo_concurrency.lockutils [req-ebf8f5ce-d221-4eeb-ae76-2101a2cd6e5f req-385de9c2-3321-41a9-847e-1a7a9197a449 service nova] Lock "1d047728-50d8-465c-b217-6cbe9bbea3e8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.013920] env[63372]: DEBUG nova.compute.manager [req-ebf8f5ce-d221-4eeb-ae76-2101a2cd6e5f req-385de9c2-3321-41a9-847e-1a7a9197a449 service nova] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] No waiting events found dispatching network-vif-plugged-a80c4949-9726-4644-967c-f72b56d42665 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 846.014132] env[63372]: WARNING nova.compute.manager [req-ebf8f5ce-d221-4eeb-ae76-2101a2cd6e5f req-385de9c2-3321-41a9-847e-1a7a9197a449 service nova] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Received unexpected event network-vif-plugged-a80c4949-9726-4644-967c-f72b56d42665 for instance with vm_state building and task_state spawning. [ 846.014302] env[63372]: DEBUG nova.compute.manager [req-ebf8f5ce-d221-4eeb-ae76-2101a2cd6e5f req-385de9c2-3321-41a9-847e-1a7a9197a449 service nova] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Received event network-changed-a80c4949-9726-4644-967c-f72b56d42665 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.014495] env[63372]: DEBUG nova.compute.manager [req-ebf8f5ce-d221-4eeb-ae76-2101a2cd6e5f req-385de9c2-3321-41a9-847e-1a7a9197a449 service nova] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Refreshing instance network info cache due to event network-changed-a80c4949-9726-4644-967c-f72b56d42665. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 846.014730] env[63372]: DEBUG oslo_concurrency.lockutils [req-ebf8f5ce-d221-4eeb-ae76-2101a2cd6e5f req-385de9c2-3321-41a9-847e-1a7a9197a449 service nova] Acquiring lock "refresh_cache-1d047728-50d8-465c-b217-6cbe9bbea3e8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.014902] env[63372]: DEBUG oslo_concurrency.lockutils [req-ebf8f5ce-d221-4eeb-ae76-2101a2cd6e5f req-385de9c2-3321-41a9-847e-1a7a9197a449 service nova] Acquired lock "refresh_cache-1d047728-50d8-465c-b217-6cbe9bbea3e8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.015159] env[63372]: DEBUG nova.network.neutron [req-ebf8f5ce-d221-4eeb-ae76-2101a2cd6e5f req-385de9c2-3321-41a9-847e-1a7a9197a449 service nova] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Refreshing network info cache for port a80c4949-9726-4644-967c-f72b56d42665 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 846.154420] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024260, 'name': ReconfigVM_Task, 'duration_secs': 0.478587} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.155095] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 07dab964-a741-4723-98a3-0cfbe4eff975/07dab964-a741-4723-98a3-0cfbe4eff975.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.155860] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fb664b4f-8f89-465d-9991-95ca8bdcfb65 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.165178] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Waiting for the task: (returnval){ [ 846.165178] env[63372]: value = "task-1024265" [ 846.165178] env[63372]: _type = "Task" [ 846.165178] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.174850] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024265, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.196870] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4a4c3008-57c3-4b47-aa02-9c062dd9dd23 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "368101ee-e4fd-4fe4-b7c8-75c6b210189f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.920s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.211651] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024262, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.239607] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024258, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.257191] env[63372]: DEBUG oslo_vmware.api [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024263, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.323309} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.257297] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.257460] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 846.257671] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 846.257885] env[63372]: INFO nova.compute.manager [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Took 1.16 seconds to destroy the instance on the hypervisor. [ 846.258194] env[63372]: DEBUG oslo.service.loopingcall [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.258431] env[63372]: DEBUG nova.compute.manager [-] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 846.258530] env[63372]: DEBUG nova.network.neutron [-] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 846.413376] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024264, 'name': CreateVM_Task, 'duration_secs': 0.463206} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.413516] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 846.417414] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.417686] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.417974] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 846.419100] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11b4fd34-6cc2-489b-a621-41636aee3172 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.424536] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 846.424536] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5269b230-749a-f4e9-1780-057e39a1384e" [ 846.424536] env[63372]: _type = "Task" [ 846.424536] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.432137] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5269b230-749a-f4e9-1780-057e39a1384e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.541080] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967ce892-8b52-4cff-872b-be371c24d8ee {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.549022] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44020c88-fc6a-4081-86ca-3a6e6a879fe2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.583296] env[63372]: DEBUG nova.network.neutron [-] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.592625] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37ad367-4fc3-48d2-9531-665494f0935c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.601066] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49bc83e7-cbe6-4023-b0a6-653b9c04ada8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.618977] env[63372]: DEBUG nova.compute.provider_tree [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.676069] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024265, 'name': Rename_Task, 'duration_secs': 0.274485} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.676446] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 846.676715] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1eeed15d-9d53-4625-a021-870af6f6da69 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.683593] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Waiting for the task: (returnval){ [ 846.683593] env[63372]: value = "task-1024266" [ 846.683593] env[63372]: _type = "Task" [ 846.683593] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.691858] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024266, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.714915] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024262, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.666232} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.715228] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 798047e4-1a2b-458c-86c2-5567197d1861/798047e4-1a2b-458c-86c2-5567197d1861.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 846.715474] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.715764] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f7c2093-fa6a-4dd9-a2f0-9cb86071f956 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.722238] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 846.722238] env[63372]: value = "task-1024267" [ 846.722238] env[63372]: _type = "Task" [ 846.722238] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.734842] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024267, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.738323] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024258, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.929821] env[63372]: DEBUG nova.network.neutron [req-ebf8f5ce-d221-4eeb-ae76-2101a2cd6e5f req-385de9c2-3321-41a9-847e-1a7a9197a449 service nova] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Updated VIF entry in instance network info cache for port a80c4949-9726-4644-967c-f72b56d42665. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 846.931180] env[63372]: DEBUG nova.network.neutron [req-ebf8f5ce-d221-4eeb-ae76-2101a2cd6e5f req-385de9c2-3321-41a9-847e-1a7a9197a449 service nova] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Updating instance_info_cache with network_info: [{"id": "a80c4949-9726-4644-967c-f72b56d42665", "address": "fa:16:3e:c5:a2:4d", "network": {"id": "49776280-f4e5-4f22-98fd-a6e6ff70c80f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1268791360-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41ca5f8268434d329deec67fb2867b59", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa80c4949-97", "ovs_interfaceid": "a80c4949-9726-4644-967c-f72b56d42665", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.938858] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5269b230-749a-f4e9-1780-057e39a1384e, 'name': SearchDatastore_Task, 'duration_secs': 0.027608} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.938858] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.938858] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.938858] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.938858] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.939156] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.939356] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c125a94e-ee47-4d21-8156-0bff12927865 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.949116] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.949116] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 846.950025] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0737bfcf-0d39-47cb-8526-66d49a483051 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.955598] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 846.955598] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52027926-5340-7cff-2811-ea175df69281" [ 846.955598] env[63372]: _type = "Task" [ 846.955598] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.965494] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52027926-5340-7cff-2811-ea175df69281, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.971183] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "ae6f3f42-7213-4ab1-b74c-1a557df6748b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.971980] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "ae6f3f42-7213-4ab1-b74c-1a557df6748b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.971980] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "ae6f3f42-7213-4ab1-b74c-1a557df6748b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.971980] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "ae6f3f42-7213-4ab1-b74c-1a557df6748b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.972381] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "ae6f3f42-7213-4ab1-b74c-1a557df6748b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.978174] env[63372]: INFO nova.compute.manager [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Terminating instance [ 846.980921] env[63372]: DEBUG nova.compute.manager [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 846.981176] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 846.982070] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4473d364-5074-49c7-a202-1dd7f0e8a9e9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.992210] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 846.992523] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1db7f1aa-45ed-450d-9491-76e96a5b7e42 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.998876] env[63372]: DEBUG oslo_vmware.api [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 846.998876] env[63372]: value = "task-1024268" [ 846.998876] env[63372]: _type = "Task" [ 846.998876] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.008106] env[63372]: DEBUG oslo_vmware.api [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024268, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.092967] env[63372]: INFO nova.compute.manager [-] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Took 1.31 seconds to deallocate network for instance. [ 847.122573] env[63372]: DEBUG nova.scheduler.client.report [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.154052] env[63372]: DEBUG nova.network.neutron [-] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.194272] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024266, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.234668] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024267, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083022} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.238191] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.238473] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024258, 'name': CloneVM_Task, 'duration_secs': 1.979294} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.239198] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a5043e-aeb0-4a34-a703-b55668df0937 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.241668] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Created linked-clone VM from snapshot [ 847.242394] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e97eee2-ba8e-4bc4-99a7-a12f17de6c5d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.266386] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 798047e4-1a2b-458c-86c2-5567197d1861/798047e4-1a2b-458c-86c2-5567197d1861.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.270440] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-125be20f-16ac-4bc7-b9b2-851a1092ff4f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.287388] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Uploading image 8988d48c-9c55-4e0b-ad32-c37a4c0ae476 {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 847.296743] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 847.296743] env[63372]: value = "task-1024269" [ 847.296743] env[63372]: _type = "Task" [ 847.296743] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.306442] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024269, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.314120] env[63372]: DEBUG oslo_vmware.rw_handles [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 847.314120] env[63372]: value = "vm-227382" [ 847.314120] env[63372]: _type = "VirtualMachine" [ 847.314120] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 847.314428] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-fdf3d47d-ba06-4c6a-b6f3-51065da1ae37 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.321383] env[63372]: DEBUG oslo_vmware.rw_handles [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lease: (returnval){ [ 847.321383] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fc147b-900e-3ca2-ea18-66954a3f6b1e" [ 847.321383] env[63372]: _type = "HttpNfcLease" [ 847.321383] env[63372]: } obtained for exporting VM: (result){ [ 847.321383] env[63372]: value = "vm-227382" [ 847.321383] env[63372]: _type = "VirtualMachine" [ 847.321383] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 847.321749] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the lease: (returnval){ [ 847.321749] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fc147b-900e-3ca2-ea18-66954a3f6b1e" [ 847.321749] env[63372]: _type = "HttpNfcLease" [ 847.321749] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 847.329265] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 847.329265] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fc147b-900e-3ca2-ea18-66954a3f6b1e" [ 847.329265] env[63372]: _type = "HttpNfcLease" [ 847.329265] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 847.433313] env[63372]: DEBUG oslo_concurrency.lockutils [req-ebf8f5ce-d221-4eeb-ae76-2101a2cd6e5f req-385de9c2-3321-41a9-847e-1a7a9197a449 service nova] Releasing lock "refresh_cache-1d047728-50d8-465c-b217-6cbe9bbea3e8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.466188] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52027926-5340-7cff-2811-ea175df69281, 'name': SearchDatastore_Task, 'duration_secs': 0.011923} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.466993] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ecd553e-201e-4d41-8780-4fac55c3de0e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.473570] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 847.473570] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]524f3878-95ce-a062-de36-0cc0077c6038" [ 847.473570] env[63372]: _type = "Task" [ 847.473570] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.481053] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524f3878-95ce-a062-de36-0cc0077c6038, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.507916] env[63372]: DEBUG oslo_vmware.api [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024268, 'name': PowerOffVM_Task, 'duration_secs': 0.212236} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.508270] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 847.508459] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 847.508704] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-672aeeed-c0c8-47dd-abfd-cf62fd3d52ff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.581983] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 847.582257] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 847.582453] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Deleting the datastore file [datastore1] ae6f3f42-7213-4ab1-b74c-1a557df6748b {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 847.582779] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11f405d2-0359-4ee7-996a-c3dd8f94c1f0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.589761] env[63372]: DEBUG oslo_vmware.api [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 847.589761] env[63372]: value = "task-1024272" [ 847.589761] env[63372]: _type = "Task" [ 847.589761] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.598512] env[63372]: DEBUG oslo_vmware.api [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024272, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.600013] env[63372]: DEBUG oslo_concurrency.lockutils [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.631189] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.988s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.633704] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.373s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.633983] env[63372]: DEBUG nova.objects.instance [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lazy-loading 'resources' on Instance uuid 357505d0-f306-4e11-8a62-e03cfab2b7c5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.659605] env[63372]: INFO nova.compute.manager [-] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Took 1.40 seconds to deallocate network for instance. [ 847.660940] env[63372]: INFO nova.scheduler.client.report [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Deleted allocations for instance 8283b736-ad02-4082-97b7-561bd5c5da93 [ 847.693853] env[63372]: DEBUG oslo_vmware.api [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024266, 'name': PowerOnVM_Task, 'duration_secs': 0.685766} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.695519] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 847.698306] env[63372]: INFO nova.compute.manager [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Took 9.28 seconds to spawn the instance on the hypervisor. [ 847.698306] env[63372]: DEBUG nova.compute.manager [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.699163] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912e2c27-4441-4652-a980-fa687952b075 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.702482] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "c0b9bd7e-92e2-44a1-b461-e54c4411ddc8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.702482] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "c0b9bd7e-92e2-44a1-b461-e54c4411ddc8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.807431] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024269, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.829657] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 847.829657] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fc147b-900e-3ca2-ea18-66954a3f6b1e" [ 847.829657] env[63372]: _type = "HttpNfcLease" [ 847.829657] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 847.830017] env[63372]: DEBUG oslo_vmware.rw_handles [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 847.830017] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fc147b-900e-3ca2-ea18-66954a3f6b1e" [ 847.830017] env[63372]: _type = "HttpNfcLease" [ 847.830017] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 847.830737] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c571d5a9-4f86-47a4-80e1-6c2020997616 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.838262] env[63372]: DEBUG oslo_vmware.rw_handles [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520aa003-24da-8b79-b377-4f476c8e516b/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 847.838435] env[63372]: DEBUG oslo_vmware.rw_handles [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520aa003-24da-8b79-b377-4f476c8e516b/disk-0.vmdk for reading. {{(pid=63372) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 847.934330] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ff213be3-bbd2-4708-8236-5259a098f508 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.983181] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524f3878-95ce-a062-de36-0cc0077c6038, 'name': SearchDatastore_Task, 'duration_secs': 0.01087} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.983463] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.983729] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 1d047728-50d8-465c-b217-6cbe9bbea3e8/1d047728-50d8-465c-b217-6cbe9bbea3e8.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 847.983987] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-14cc970d-d217-43d5-9b97-fe90521e46d1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.990053] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 847.990053] env[63372]: value = "task-1024273" [ 847.990053] env[63372]: _type = "Task" [ 847.990053] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.997879] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024273, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.043597] env[63372]: DEBUG nova.compute.manager [req-fe7d71de-316e-48ae-8491-f3f72299a05b req-69338a9c-4c35-40ea-a6aa-ffc4051f97c6 service nova] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Received event network-vif-deleted-5bf209d9-9f64-410a-b14f-19cd792d4f7c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.043821] env[63372]: DEBUG nova.compute.manager [req-fe7d71de-316e-48ae-8491-f3f72299a05b req-69338a9c-4c35-40ea-a6aa-ffc4051f97c6 service nova] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Received event network-vif-deleted-92762aef-8192-454f-95dc-1eb8d39cf9a9 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.100262] env[63372]: DEBUG oslo_vmware.api [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024272, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14467} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.100561] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 848.100767] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 848.100950] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 848.101183] env[63372]: INFO nova.compute.manager [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 848.101619] env[63372]: DEBUG oslo.service.loopingcall [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.101772] env[63372]: DEBUG nova.compute.manager [-] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 848.101772] env[63372]: DEBUG nova.network.neutron [-] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 848.174854] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.175559] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a5345d1-af98-4b56-baa0-d5025b8e54d9 tempest-MigrationsAdminTest-1076957705 tempest-MigrationsAdminTest-1076957705-project-member] Lock "8283b736-ad02-4082-97b7-561bd5c5da93" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.403s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.205274] env[63372]: DEBUG nova.compute.manager [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 848.220054] env[63372]: INFO nova.compute.manager [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Took 25.57 seconds to build instance. [ 848.311320] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024269, 'name': ReconfigVM_Task, 'duration_secs': 0.6511} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.311744] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 798047e4-1a2b-458c-86c2-5567197d1861/798047e4-1a2b-458c-86c2-5567197d1861.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.312513] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-07cfd02b-f72b-4d8c-b67e-254e424c6bbc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.321912] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 848.321912] env[63372]: value = "task-1024274" [ 848.321912] env[63372]: _type = "Task" [ 848.321912] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.334149] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024274, 'name': Rename_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.504352] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024273, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.515533] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc2d5e9-970c-42b8-8ec0-da4cf5453fa7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.524961] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ba533e-e6ce-4a03-a121-f2c18f7d8cc2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.565933] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff0baee-6b73-4ee7-a5f6-cc4c386c2e03 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.575691] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d34e2e59-f150-42b9-8058-24db5ecb8dcf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.591913] env[63372]: DEBUG nova.compute.provider_tree [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.722614] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c177136-e39f-424c-a18c-289ef2cc490f tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Lock "07dab964-a741-4723-98a3-0cfbe4eff975" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.644s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.730813] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.801562] env[63372]: DEBUG nova.compute.manager [req-cc635e16-6348-422c-b7d4-9551bbb8b21d req-19a85862-3d00-4769-af25-9052184b612a service nova] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Received event network-vif-deleted-18da52d3-9658-4af9-bd78-16e5ab3be9d9 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.802436] env[63372]: INFO nova.compute.manager [req-cc635e16-6348-422c-b7d4-9551bbb8b21d req-19a85862-3d00-4769-af25-9052184b612a service nova] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Neutron deleted interface 18da52d3-9658-4af9-bd78-16e5ab3be9d9; detaching it from the instance and deleting it from the info cache [ 848.802436] env[63372]: DEBUG nova.network.neutron [req-cc635e16-6348-422c-b7d4-9551bbb8b21d req-19a85862-3d00-4769-af25-9052184b612a service nova] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.835362] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024274, 'name': Rename_Task, 'duration_secs': 0.300826} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.835768] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 848.836110] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40547572-ae2b-472b-a42b-907b3d139991 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.845079] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 848.845079] env[63372]: value = "task-1024275" [ 848.845079] env[63372]: _type = "Task" [ 848.845079] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.855075] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024275, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.003083] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024273, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.58384} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.003083] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 1d047728-50d8-465c-b217-6cbe9bbea3e8/1d047728-50d8-465c-b217-6cbe9bbea3e8.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 849.003986] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.005782] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-47f4d2ec-536b-4807-8891-3c2b7bd2c12a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.012507] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 849.012507] env[63372]: value = "task-1024276" [ 849.012507] env[63372]: _type = "Task" [ 849.012507] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.024213] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024276, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.095374] env[63372]: DEBUG nova.scheduler.client.report [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.099657] env[63372]: DEBUG nova.network.neutron [-] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.305768] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bcf75e0f-b3d6-4a72-b625-f1188847991c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.317023] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d173084d-aa11-4804-bb71-4c85c6473de0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.351695] env[63372]: DEBUG nova.compute.manager [req-cc635e16-6348-422c-b7d4-9551bbb8b21d req-19a85862-3d00-4769-af25-9052184b612a service nova] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Detach interface failed, port_id=18da52d3-9658-4af9-bd78-16e5ab3be9d9, reason: Instance ae6f3f42-7213-4ab1-b74c-1a557df6748b could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 849.361252] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024275, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.437618] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Acquiring lock "07dab964-a741-4723-98a3-0cfbe4eff975" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.438409] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Lock "07dab964-a741-4723-98a3-0cfbe4eff975" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.438409] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Acquiring lock "07dab964-a741-4723-98a3-0cfbe4eff975-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.438724] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Lock "07dab964-a741-4723-98a3-0cfbe4eff975-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.438724] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Lock "07dab964-a741-4723-98a3-0cfbe4eff975-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.443859] env[63372]: INFO nova.compute.manager [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Terminating instance [ 849.449511] env[63372]: DEBUG nova.compute.manager [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 849.449734] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 849.450644] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9966c788-7d58-4ec1-ae8c-cf7646228e50 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.462028] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 849.463096] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fae85a4a-eff6-49b6-a33e-a02f55196f2c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.475245] env[63372]: DEBUG oslo_vmware.api [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Waiting for the task: (returnval){ [ 849.475245] env[63372]: value = "task-1024277" [ 849.475245] env[63372]: _type = "Task" [ 849.475245] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.484668] env[63372]: DEBUG oslo_vmware.api [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024277, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.486933] env[63372]: DEBUG oslo_vmware.rw_handles [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f900f0-cd4e-45e0-f03b-6146fd73591c/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 849.487870] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ba5073-18d4-4fc7-b0b9-6e4e22b32685 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.493873] env[63372]: DEBUG oslo_vmware.rw_handles [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f900f0-cd4e-45e0-f03b-6146fd73591c/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 849.494112] env[63372]: ERROR oslo_vmware.rw_handles [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f900f0-cd4e-45e0-f03b-6146fd73591c/disk-0.vmdk due to incomplete transfer. [ 849.494463] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5ecf98c0-627e-4498-8fb7-5c4b46e23827 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.501862] env[63372]: DEBUG oslo_vmware.rw_handles [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f900f0-cd4e-45e0-f03b-6146fd73591c/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 849.502147] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Uploaded image 217465cd-6c37-4aef-bacb-fbc93077f64b to the Glance image server {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 849.504525] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 849.504865] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2598f4bb-8b82-4caf-8e81-a55d5c21ceb7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.511141] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 849.511141] env[63372]: value = "task-1024278" [ 849.511141] env[63372]: _type = "Task" [ 849.511141] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.529641] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024276, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08153} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.529887] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024278, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.530235] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 849.531169] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651133fb-1864-453d-957c-f1388fa48af8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.556026] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 1d047728-50d8-465c-b217-6cbe9bbea3e8/1d047728-50d8-465c-b217-6cbe9bbea3e8.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.556328] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5008288-73f8-4d3d-806f-cf9625546784 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.578991] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 849.578991] env[63372]: value = "task-1024279" [ 849.578991] env[63372]: _type = "Task" [ 849.578991] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.592090] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024279, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.602828] env[63372]: INFO nova.compute.manager [-] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Took 1.50 seconds to deallocate network for instance. [ 849.603711] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.970s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.608073] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.491s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.608454] env[63372]: DEBUG nova.objects.instance [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lazy-loading 'resources' on Instance uuid b182294d-2de8-4189-af7f-3e2d2c604a8b {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.634746] env[63372]: INFO nova.scheduler.client.report [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Deleted allocations for instance 357505d0-f306-4e11-8a62-e03cfab2b7c5 [ 849.863857] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024275, 'name': PowerOnVM_Task, 'duration_secs': 0.765922} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.864235] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 849.864524] env[63372]: INFO nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Took 8.91 seconds to spawn the instance on the hypervisor. [ 849.865292] env[63372]: DEBUG nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.866079] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d130f45-8957-40aa-93a0-9ed37b98851e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.987045] env[63372]: DEBUG oslo_vmware.api [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024277, 'name': PowerOffVM_Task, 'duration_secs': 0.306604} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.987529] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 849.987749] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 849.988051] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-379af335-a3bb-4d08-a84f-cf9e0499a52e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.021900] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024278, 'name': Destroy_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.069978] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 850.071169] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 850.071169] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Deleting the datastore file [datastore2] 07dab964-a741-4723-98a3-0cfbe4eff975 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 850.071169] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-111c8c37-7f86-4726-af1d-c60cc7f595dc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.078007] env[63372]: DEBUG oslo_vmware.api [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Waiting for the task: (returnval){ [ 850.078007] env[63372]: value = "task-1024281" [ 850.078007] env[63372]: _type = "Task" [ 850.078007] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.091384] env[63372]: DEBUG oslo_vmware.api [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024281, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.095510] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024279, 'name': ReconfigVM_Task, 'duration_secs': 0.335991} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.095807] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 1d047728-50d8-465c-b217-6cbe9bbea3e8/1d047728-50d8-465c-b217-6cbe9bbea3e8.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.096522] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c196ee0c-20f5-4383-a2a0-9be1aedb2dbb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.102540] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 850.102540] env[63372]: value = "task-1024282" [ 850.102540] env[63372]: _type = "Task" [ 850.102540] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.113799] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024282, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.118284] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.144661] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c4d6e809-3029-4ad8-bd0a-e1067540b979 tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "357505d0-f306-4e11-8a62-e03cfab2b7c5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.902s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.288275] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "cd0c01ac-602b-44a3-8099-84b8a50b2449" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.288577] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "cd0c01ac-602b-44a3-8099-84b8a50b2449" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.289365] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "cd0c01ac-602b-44a3-8099-84b8a50b2449-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.289651] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "cd0c01ac-602b-44a3-8099-84b8a50b2449-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.289901] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "cd0c01ac-602b-44a3-8099-84b8a50b2449-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.298539] env[63372]: INFO nova.compute.manager [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Terminating instance [ 850.301926] env[63372]: DEBUG nova.compute.manager [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 850.302221] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 850.303837] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e50c4127-adfc-4069-ab78-3bf0a2f93248 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.314684] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 850.317930] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70799298-578f-420a-b1b2-5896c3899526 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.326954] env[63372]: DEBUG oslo_vmware.api [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 850.326954] env[63372]: value = "task-1024283" [ 850.326954] env[63372]: _type = "Task" [ 850.326954] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.337874] env[63372]: DEBUG oslo_vmware.api [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024283, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.388389] env[63372]: INFO nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Took 27.20 seconds to build instance. [ 850.483620] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b17e57bf-dc21-4556-87ec-4773149edbe9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.492976] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9512ab95-0796-4aeb-bbd9-c2407ff21cb1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.539305] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d974bc-92f0-4c2e-b70b-be8983074755 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.548832] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024278, 'name': Destroy_Task, 'duration_secs': 0.688396} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.551198] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Destroyed the VM [ 850.551484] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 850.551968] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-07c68132-75eb-4883-aada-c1ddaf24194f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.554611] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2e1811-a0b4-400a-9139-2d857a3ba1da {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.574562] env[63372]: DEBUG nova.compute.provider_tree [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.577489] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 850.577489] env[63372]: value = "task-1024284" [ 850.577489] env[63372]: _type = "Task" [ 850.577489] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.591879] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024284, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.595429] env[63372]: DEBUG oslo_vmware.api [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Task: {'id': task-1024281, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.223879} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.596134] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 850.596420] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 850.596587] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 850.596771] env[63372]: INFO nova.compute.manager [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Took 1.15 seconds to destroy the instance on the hypervisor. [ 850.597067] env[63372]: DEBUG oslo.service.loopingcall [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.597353] env[63372]: DEBUG nova.compute.manager [-] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 850.597426] env[63372]: DEBUG nova.network.neutron [-] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 850.615151] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024282, 'name': Rename_Task, 'duration_secs': 0.166776} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.615461] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 850.615722] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a85ec69d-9b20-40d3-aad7-1837b26235bc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.625736] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 850.625736] env[63372]: value = "task-1024285" [ 850.625736] env[63372]: _type = "Task" [ 850.625736] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.636777] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024285, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.842712] env[63372]: DEBUG oslo_vmware.api [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024283, 'name': PowerOffVM_Task, 'duration_secs': 0.315992} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.843455] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 850.843838] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 850.844331] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2892d6ce-c696-43b8-b35b-17a84596f048 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.891863] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "798047e4-1a2b-458c-86c2-5567197d1861" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.721s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.059622] env[63372]: DEBUG nova.compute.manager [req-6ec3e35d-6f97-41e6-a5a2-e78826c6d7bb req-1f7166c3-9df9-478f-8d93-83d9faafad24 service nova] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Received event network-vif-deleted-d27ee5ea-fd81-4901-bdf2-9876321872b5 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.059622] env[63372]: INFO nova.compute.manager [req-6ec3e35d-6f97-41e6-a5a2-e78826c6d7bb req-1f7166c3-9df9-478f-8d93-83d9faafad24 service nova] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Neutron deleted interface d27ee5ea-fd81-4901-bdf2-9876321872b5; detaching it from the instance and deleting it from the info cache [ 851.063434] env[63372]: DEBUG nova.network.neutron [req-6ec3e35d-6f97-41e6-a5a2-e78826c6d7bb req-1f7166c3-9df9-478f-8d93-83d9faafad24 service nova] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.085868] env[63372]: DEBUG nova.scheduler.client.report [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.097656] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024284, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.137976] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024285, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.452670] env[63372]: DEBUG nova.network.neutron [-] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.566869] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0fe4254-584a-4eb4-8a7f-6f89828e1bf9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.582759] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9d1c63c-8f36-474e-a1e6-73f76109c6fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.597985] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.990s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.604804] env[63372]: DEBUG oslo_concurrency.lockutils [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.372s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.605073] env[63372]: DEBUG nova.objects.instance [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lazy-loading 'resources' on Instance uuid 3f66eda6-f5e9-4527-9711-849a01702580 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.612481] env[63372]: DEBUG oslo_vmware.api [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024284, 'name': RemoveSnapshot_Task, 'duration_secs': 0.785264} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.612757] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 851.612990] env[63372]: INFO nova.compute.manager [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Took 16.32 seconds to snapshot the instance on the hypervisor. [ 851.628822] env[63372]: DEBUG nova.compute.manager [req-6ec3e35d-6f97-41e6-a5a2-e78826c6d7bb req-1f7166c3-9df9-478f-8d93-83d9faafad24 service nova] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Detach interface failed, port_id=d27ee5ea-fd81-4901-bdf2-9876321872b5, reason: Instance 07dab964-a741-4723-98a3-0cfbe4eff975 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 851.630772] env[63372]: INFO nova.scheduler.client.report [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Deleted allocations for instance b182294d-2de8-4189-af7f-3e2d2c604a8b [ 851.644126] env[63372]: DEBUG oslo_vmware.api [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024285, 'name': PowerOnVM_Task, 'duration_secs': 0.63676} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.644126] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 851.644126] env[63372]: INFO nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Took 8.06 seconds to spawn the instance on the hypervisor. [ 851.644126] env[63372]: DEBUG nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.644377] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f89a20-8ba1-4dea-8b3c-d7c161f4d190 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.955979] env[63372]: INFO nova.compute.manager [-] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Took 1.36 seconds to deallocate network for instance. [ 851.976875] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 851.977162] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 851.977462] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Deleting the datastore file [datastore1] cd0c01ac-602b-44a3-8099-84b8a50b2449 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 851.977639] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ed4dbd2-a22f-4e82-b0f9-b08624ca3b14 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.987027] env[63372]: DEBUG oslo_vmware.api [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for the task: (returnval){ [ 851.987027] env[63372]: value = "task-1024287" [ 851.987027] env[63372]: _type = "Task" [ 851.987027] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.000574] env[63372]: DEBUG oslo_vmware.api [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024287, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.143394] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1c4d7716-d808-4d53-8fbe-7937499c94c9 tempest-ListImageFiltersTestJSON-523138203 tempest-ListImageFiltersTestJSON-523138203-project-member] Lock "b182294d-2de8-4189-af7f-3e2d2c604a8b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.588s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.176160] env[63372]: INFO nova.compute.manager [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Took 28.99 seconds to build instance. [ 852.181955] env[63372]: DEBUG nova.compute.manager [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Found 3 images (rotation: 2) {{(pid=63372) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 852.182309] env[63372]: DEBUG nova.compute.manager [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Rotating out 1 backups {{(pid=63372) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 852.182513] env[63372]: DEBUG nova.compute.manager [None req-4b1d81a4-92c8-4ce3-b192-fa2053f58607 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Deleting image de1edf71-70ee-4a4d-9486-cbfcce731c28 {{(pid=63372) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 852.464901] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3204499a-2109-4ecf-9618-8f924f9c21ab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.471459] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.475357] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023a2298-0e67-4871-a880-5bc980b3454e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.515573] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab70c76c-13b4-42ba-9b16-ac71159e6c6b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.526819] env[63372]: DEBUG oslo_vmware.api [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Task: {'id': task-1024287, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.258772} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.528162] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0d892d-2fc8-48df-bf59-3d3488e0f8f7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.532138] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 852.532372] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 852.532565] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 852.532757] env[63372]: INFO nova.compute.manager [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Took 2.23 seconds to destroy the instance on the hypervisor. [ 852.533029] env[63372]: DEBUG oslo.service.loopingcall [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.533256] env[63372]: DEBUG nova.compute.manager [-] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 852.533357] env[63372]: DEBUG nova.network.neutron [-] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 852.545459] env[63372]: DEBUG nova.compute.provider_tree [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.679065] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5a85c95a-2ded-4163-ae0e-0b53ec007713 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "1d047728-50d8-465c-b217-6cbe9bbea3e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.506s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.048482] env[63372]: DEBUG nova.scheduler.client.report [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.401162] env[63372]: DEBUG nova.compute.manager [req-dcf7afe1-b376-430c-ba45-735bff798db4 req-67760777-56a4-46bd-ac64-afc89c5ebd9f service nova] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Received event network-vif-deleted-263cc51f-62e9-4a41-9d93-6f8c90cb320d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.401423] env[63372]: INFO nova.compute.manager [req-dcf7afe1-b376-430c-ba45-735bff798db4 req-67760777-56a4-46bd-ac64-afc89c5ebd9f service nova] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Neutron deleted interface 263cc51f-62e9-4a41-9d93-6f8c90cb320d; detaching it from the instance and deleting it from the info cache [ 853.401609] env[63372]: DEBUG nova.network.neutron [req-dcf7afe1-b376-430c-ba45-735bff798db4 req-67760777-56a4-46bd-ac64-afc89c5ebd9f service nova] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.498914] env[63372]: DEBUG nova.network.neutron [-] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.555111] env[63372]: DEBUG oslo_concurrency.lockutils [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.950s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.559073] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.808s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.564835] env[63372]: INFO nova.compute.claims [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 853.592036] env[63372]: INFO nova.scheduler.client.report [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Deleted allocations for instance 3f66eda6-f5e9-4527-9711-849a01702580 [ 853.910052] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8cd5b4e0-0570-449b-8e53-b2d19a625499 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.921918] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2fd827e-1155-4586-9e29-1c944fab2f0c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.978118] env[63372]: DEBUG nova.compute.manager [req-dcf7afe1-b376-430c-ba45-735bff798db4 req-67760777-56a4-46bd-ac64-afc89c5ebd9f service nova] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Detach interface failed, port_id=263cc51f-62e9-4a41-9d93-6f8c90cb320d, reason: Instance cd0c01ac-602b-44a3-8099-84b8a50b2449 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 854.005093] env[63372]: INFO nova.compute.manager [-] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Took 1.47 seconds to deallocate network for instance. [ 854.082991] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "798047e4-1a2b-458c-86c2-5567197d1861" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.082991] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "798047e4-1a2b-458c-86c2-5567197d1861" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.082991] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "798047e4-1a2b-458c-86c2-5567197d1861-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.082991] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "798047e4-1a2b-458c-86c2-5567197d1861-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.082991] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "798047e4-1a2b-458c-86c2-5567197d1861-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.087207] env[63372]: INFO nova.compute.manager [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Terminating instance [ 854.093329] env[63372]: DEBUG nova.compute.manager [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 854.093581] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 854.095367] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bbdd5ef-4b6a-446b-847f-39eabfc3ca40 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.103115] env[63372]: DEBUG oslo_concurrency.lockutils [None req-652a87e4-5d74-49e2-a61a-d495e21d3608 tempest-ServerShowV247Test-1114235032 tempest-ServerShowV247Test-1114235032-project-member] Lock "3f66eda6-f5e9-4527-9711-849a01702580" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.626s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.108980] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 854.109513] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a9a674c2-1c05-42ac-8fdd-5fc74a4d0739 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.118362] env[63372]: DEBUG oslo_vmware.api [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 854.118362] env[63372]: value = "task-1024288" [ 854.118362] env[63372]: _type = "Task" [ 854.118362] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.130631] env[63372]: DEBUG oslo_vmware.api [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024288, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.515236] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.603323] env[63372]: DEBUG oslo_concurrency.lockutils [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "1d047728-50d8-465c-b217-6cbe9bbea3e8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.603323] env[63372]: DEBUG oslo_concurrency.lockutils [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "1d047728-50d8-465c-b217-6cbe9bbea3e8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.603323] env[63372]: DEBUG oslo_concurrency.lockutils [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "1d047728-50d8-465c-b217-6cbe9bbea3e8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.603489] env[63372]: DEBUG oslo_concurrency.lockutils [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "1d047728-50d8-465c-b217-6cbe9bbea3e8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.603585] env[63372]: DEBUG oslo_concurrency.lockutils [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "1d047728-50d8-465c-b217-6cbe9bbea3e8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.606836] env[63372]: INFO nova.compute.manager [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Terminating instance [ 854.609195] env[63372]: DEBUG nova.compute.manager [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 854.609433] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 854.610344] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d57907-48f4-4ea0-bb98-a1d5f21612be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.627492] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 854.627888] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-133e11ad-69d6-4ae5-937c-5b02acab21a3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.636370] env[63372]: DEBUG oslo_vmware.api [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024288, 'name': PowerOffVM_Task, 'duration_secs': 0.257779} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.640731] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 854.640961] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 854.641711] env[63372]: DEBUG oslo_vmware.api [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 854.641711] env[63372]: value = "task-1024289" [ 854.641711] env[63372]: _type = "Task" [ 854.641711] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.643068] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05f5f47e-ab71-4f34-ab3c-02c87d83d67b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.656340] env[63372]: DEBUG oslo_vmware.api [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024289, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.733230] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 854.733657] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 854.733917] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Deleting the datastore file [datastore2] 798047e4-1a2b-458c-86c2-5567197d1861 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.734296] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-818ad94e-781b-4734-96d7-f7bf21ff3b51 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.742885] env[63372]: DEBUG oslo_vmware.api [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 854.742885] env[63372]: value = "task-1024291" [ 854.742885] env[63372]: _type = "Task" [ 854.742885] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.760957] env[63372]: DEBUG oslo_vmware.api [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024291, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.962205] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea664358-1d8c-43a1-974c-b4af91d85381 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.973666] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4e4d9b-c80e-4f62-9836-96ccf6e3bc93 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.010520] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5acfe530-044a-4b55-b038-b9b9d162d4fc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.019272] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd582912-3aab-4b26-8c6e-96e1201eb398 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.033988] env[63372]: DEBUG nova.compute.provider_tree [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.156841] env[63372]: DEBUG oslo_vmware.api [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024289, 'name': PowerOffVM_Task, 'duration_secs': 0.241627} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.157040] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 855.157081] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 855.157338] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2c6c3310-e763-4acb-88cf-b793e51a1960 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.234956] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 855.235298] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 855.235549] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Deleting the datastore file [datastore2] 1d047728-50d8-465c-b217-6cbe9bbea3e8 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 855.235884] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0fcc39b0-1b63-46fa-96a9-987f9afd4b3f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.245566] env[63372]: DEBUG oslo_vmware.api [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for the task: (returnval){ [ 855.245566] env[63372]: value = "task-1024293" [ 855.245566] env[63372]: _type = "Task" [ 855.245566] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.259659] env[63372]: DEBUG oslo_vmware.api [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024291, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.21875} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.263765] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 855.264106] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 855.264364] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 855.264790] env[63372]: INFO nova.compute.manager [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Took 1.17 seconds to destroy the instance on the hypervisor. [ 855.265220] env[63372]: DEBUG oslo.service.loopingcall [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.265527] env[63372]: DEBUG oslo_vmware.api [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024293, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.265800] env[63372]: DEBUG nova.compute.manager [-] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 855.265943] env[63372]: DEBUG nova.network.neutron [-] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 855.540131] env[63372]: DEBUG nova.scheduler.client.report [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.764058] env[63372]: DEBUG oslo_vmware.api [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Task: {'id': task-1024293, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.349997} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.764058] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 855.764058] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 855.764058] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 855.764058] env[63372]: INFO nova.compute.manager [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 855.764058] env[63372]: DEBUG oslo.service.loopingcall [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.764058] env[63372]: DEBUG nova.compute.manager [-] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 855.764058] env[63372]: DEBUG nova.network.neutron [-] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 856.043079] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.043588] env[63372]: DEBUG nova.compute.manager [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 856.046444] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.459s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.046696] env[63372]: DEBUG nova.objects.instance [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Lazy-loading 'resources' on Instance uuid cf673ac1-2c7d-468b-83ec-c723d5182457 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.349451] env[63372]: DEBUG nova.network.neutron [-] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.431673] env[63372]: DEBUG nova.compute.manager [req-c94a8ba1-d1f2-42d6-839e-16dfa041a7b8 req-09076dd0-6350-4848-9b55-72c13dcfc801 service nova] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Received event network-vif-deleted-a80c4949-9726-4644-967c-f72b56d42665 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.431913] env[63372]: INFO nova.compute.manager [req-c94a8ba1-d1f2-42d6-839e-16dfa041a7b8 req-09076dd0-6350-4848-9b55-72c13dcfc801 service nova] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Neutron deleted interface a80c4949-9726-4644-967c-f72b56d42665; detaching it from the instance and deleting it from the info cache [ 856.432271] env[63372]: DEBUG nova.network.neutron [req-c94a8ba1-d1f2-42d6-839e-16dfa041a7b8 req-09076dd0-6350-4848-9b55-72c13dcfc801 service nova] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.504069] env[63372]: DEBUG nova.network.neutron [-] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.512905] env[63372]: DEBUG nova.compute.manager [req-fb56a689-06af-49d3-a032-08c50206a5e7 req-0ce40cce-1c08-483e-b5ce-e775916665f6 service nova] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Received event network-vif-deleted-bfc2f0a4-c40a-49e2-9d87-e7f9593a583a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.513160] env[63372]: INFO nova.compute.manager [req-fb56a689-06af-49d3-a032-08c50206a5e7 req-0ce40cce-1c08-483e-b5ce-e775916665f6 service nova] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Neutron deleted interface bfc2f0a4-c40a-49e2-9d87-e7f9593a583a; detaching it from the instance and deleting it from the info cache [ 856.513357] env[63372]: DEBUG nova.network.neutron [req-fb56a689-06af-49d3-a032-08c50206a5e7 req-0ce40cce-1c08-483e-b5ce-e775916665f6 service nova] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.550236] env[63372]: DEBUG nova.compute.utils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 856.555554] env[63372]: DEBUG nova.compute.manager [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 856.555661] env[63372]: DEBUG nova.network.neutron [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 856.609144] env[63372]: DEBUG nova.policy [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7ffe6776b664c00b9086046f9171ec6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46d2cb188a004dc385f2aa15458be929', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 856.838252] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2752af63-1509-487b-975d-98c7a6f78712 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.846900] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c480a0c-3fac-4c7a-8ed5-d6f063b961f6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.853831] env[63372]: INFO nova.compute.manager [-] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Took 1.59 seconds to deallocate network for instance. [ 856.893566] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b77afe-34df-4de5-b2ae-ea57dc3150b5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.902586] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3a1078-f76a-45d5-8ec3-1ade1a94a796 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.921021] env[63372]: DEBUG nova.compute.provider_tree [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.921021] env[63372]: DEBUG nova.network.neutron [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Successfully created port: b02b2fb0-9262-4f24-a677-c4f61362410a {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 856.936117] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-96349ac9-033f-4216-a7fc-e967a2dfa10e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.948147] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a032d9d8-f90d-43b5-be09-1a81fa84c3a8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.978990] env[63372]: DEBUG nova.compute.manager [req-c94a8ba1-d1f2-42d6-839e-16dfa041a7b8 req-09076dd0-6350-4848-9b55-72c13dcfc801 service nova] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Detach interface failed, port_id=a80c4949-9726-4644-967c-f72b56d42665, reason: Instance 1d047728-50d8-465c-b217-6cbe9bbea3e8 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 857.007401] env[63372]: INFO nova.compute.manager [-] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Took 1.24 seconds to deallocate network for instance. [ 857.016255] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eeecdb61-1960-4d15-a5de-c2ba1361c032 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.027124] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec0db2ce-42af-4b68-ac94-f153f20559fd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.058628] env[63372]: DEBUG nova.compute.manager [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 857.066116] env[63372]: DEBUG nova.compute.manager [req-fb56a689-06af-49d3-a032-08c50206a5e7 req-0ce40cce-1c08-483e-b5ce-e775916665f6 service nova] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Detach interface failed, port_id=bfc2f0a4-c40a-49e2-9d87-e7f9593a583a, reason: Instance 798047e4-1a2b-458c-86c2-5567197d1861 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 857.397653] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.423882] env[63372]: DEBUG nova.scheduler.client.report [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.520036] env[63372]: DEBUG oslo_concurrency.lockutils [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.701901] env[63372]: DEBUG oslo_vmware.rw_handles [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520aa003-24da-8b79-b377-4f476c8e516b/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 857.704167] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2aa8774-ef69-46b3-9009-92d4ba3259aa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.709667] env[63372]: DEBUG oslo_vmware.rw_handles [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520aa003-24da-8b79-b377-4f476c8e516b/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 857.709831] env[63372]: ERROR oslo_vmware.rw_handles [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520aa003-24da-8b79-b377-4f476c8e516b/disk-0.vmdk due to incomplete transfer. [ 857.710092] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e4fa9d7e-d943-4dca-80bc-833330d46c75 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.717513] env[63372]: DEBUG oslo_vmware.rw_handles [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520aa003-24da-8b79-b377-4f476c8e516b/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 857.717717] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Uploaded image 8988d48c-9c55-4e0b-ad32-c37a4c0ae476 to the Glance image server {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 857.719590] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 857.720011] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ffc84188-0142-4c83-89bb-1f57d4565936 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.727471] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 857.727471] env[63372]: value = "task-1024294" [ 857.727471] env[63372]: _type = "Task" [ 857.727471] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.735884] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024294, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.806232] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b001fdb5-9a35-4760-a137-3bfbbcb4c544 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.806654] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b001fdb5-9a35-4760-a137-3bfbbcb4c544 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.806876] env[63372]: DEBUG nova.compute.manager [None req-b001fdb5-9a35-4760-a137-3bfbbcb4c544 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.807825] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7789654-c44f-425b-86e0-94531a55229d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.815019] env[63372]: DEBUG nova.compute.manager [None req-b001fdb5-9a35-4760-a137-3bfbbcb4c544 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63372) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 857.815623] env[63372]: DEBUG nova.objects.instance [None req-b001fdb5-9a35-4760-a137-3bfbbcb4c544 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'flavor' on Instance uuid 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.929689] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.883s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.932356] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.890s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.933907] env[63372]: INFO nova.compute.claims [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.956826] env[63372]: INFO nova.scheduler.client.report [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Deleted allocations for instance cf673ac1-2c7d-468b-83ec-c723d5182457 [ 858.073491] env[63372]: DEBUG nova.compute.manager [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 858.096729] env[63372]: DEBUG nova.virt.hardware [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.097054] env[63372]: DEBUG nova.virt.hardware [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.097196] env[63372]: DEBUG nova.virt.hardware [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.097409] env[63372]: DEBUG nova.virt.hardware [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.097572] env[63372]: DEBUG nova.virt.hardware [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.097795] env[63372]: DEBUG nova.virt.hardware [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.098256] env[63372]: DEBUG nova.virt.hardware [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.098256] env[63372]: DEBUG nova.virt.hardware [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.098480] env[63372]: DEBUG nova.virt.hardware [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.098521] env[63372]: DEBUG nova.virt.hardware [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.098705] env[63372]: DEBUG nova.virt.hardware [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.099669] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aba2ea9-cb54-4e42-80b5-090d846fd2e7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.108789] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d5aa11-faf8-40c6-91ac-01aa78b80208 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.238545] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024294, 'name': Destroy_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.321485] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b001fdb5-9a35-4760-a137-3bfbbcb4c544 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 858.321712] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe264c84-cb88-469d-a6e1-0c97e3a31670 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.330535] env[63372]: DEBUG oslo_vmware.api [None req-b001fdb5-9a35-4760-a137-3bfbbcb4c544 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 858.330535] env[63372]: value = "task-1024295" [ 858.330535] env[63372]: _type = "Task" [ 858.330535] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.340455] env[63372]: DEBUG oslo_vmware.api [None req-b001fdb5-9a35-4760-a137-3bfbbcb4c544 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024295, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.464731] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f42a8856-50b3-4559-95f2-73c5a74b2eb2 tempest-ServersTestManualDisk-806885957 tempest-ServersTestManualDisk-806885957-project-member] Lock "cf673ac1-2c7d-468b-83ec-c723d5182457" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.248s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.481423] env[63372]: DEBUG nova.network.neutron [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Successfully updated port: b02b2fb0-9262-4f24-a677-c4f61362410a {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 858.532270] env[63372]: DEBUG nova.compute.manager [req-21be154f-c2aa-4b95-a8aa-f46303943570 req-45027ca0-4ca3-4ade-9628-47ebcee796e1 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Received event network-vif-plugged-b02b2fb0-9262-4f24-a677-c4f61362410a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.532531] env[63372]: DEBUG oslo_concurrency.lockutils [req-21be154f-c2aa-4b95-a8aa-f46303943570 req-45027ca0-4ca3-4ade-9628-47ebcee796e1 service nova] Acquiring lock "13bde897-8446-42a2-b02d-2f5b48e6f432-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.532752] env[63372]: DEBUG oslo_concurrency.lockutils [req-21be154f-c2aa-4b95-a8aa-f46303943570 req-45027ca0-4ca3-4ade-9628-47ebcee796e1 service nova] Lock "13bde897-8446-42a2-b02d-2f5b48e6f432-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.532929] env[63372]: DEBUG oslo_concurrency.lockutils [req-21be154f-c2aa-4b95-a8aa-f46303943570 req-45027ca0-4ca3-4ade-9628-47ebcee796e1 service nova] Lock "13bde897-8446-42a2-b02d-2f5b48e6f432-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.533154] env[63372]: DEBUG nova.compute.manager [req-21be154f-c2aa-4b95-a8aa-f46303943570 req-45027ca0-4ca3-4ade-9628-47ebcee796e1 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] No waiting events found dispatching network-vif-plugged-b02b2fb0-9262-4f24-a677-c4f61362410a {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 858.533331] env[63372]: WARNING nova.compute.manager [req-21be154f-c2aa-4b95-a8aa-f46303943570 req-45027ca0-4ca3-4ade-9628-47ebcee796e1 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Received unexpected event network-vif-plugged-b02b2fb0-9262-4f24-a677-c4f61362410a for instance with vm_state building and task_state spawning. [ 858.740723] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024294, 'name': Destroy_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.841017] env[63372]: DEBUG oslo_vmware.api [None req-b001fdb5-9a35-4760-a137-3bfbbcb4c544 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024295, 'name': PowerOffVM_Task, 'duration_secs': 0.339062} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.841484] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b001fdb5-9a35-4760-a137-3bfbbcb4c544 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 858.841677] env[63372]: DEBUG nova.compute.manager [None req-b001fdb5-9a35-4760-a137-3bfbbcb4c544 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.842438] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392e8c1f-21bb-47e5-8cc7-55b42cbedf6e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.984147] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.984530] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.984704] env[63372]: DEBUG nova.network.neutron [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 859.216533] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8719d5f5-65b1-4307-a553-703cc8dd9860 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.224640] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1216c7c3-0dac-4aad-bbf9-dbd16a5cef60 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.262847] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b95e285-a3f1-4c08-b93c-3ddbe10b8b8d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.268589] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024294, 'name': Destroy_Task, 'duration_secs': 1.382547} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.269257] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Destroyed the VM [ 859.269555] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 859.269823] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c9fbc64f-d5af-4d93-a92c-f6be52b0d0a0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.279220] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc27b3c-0b47-4ca4-8831-3b0bf9b7f86f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.280524] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 859.280524] env[63372]: value = "task-1024296" [ 859.280524] env[63372]: _type = "Task" [ 859.280524] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.289714] env[63372]: DEBUG nova.compute.provider_tree [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.303295] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024296, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.356094] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b001fdb5-9a35-4760-a137-3bfbbcb4c544 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.549s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.532120] env[63372]: DEBUG nova.network.neutron [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 859.753316] env[63372]: DEBUG nova.network.neutron [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Updating instance_info_cache with network_info: [{"id": "b02b2fb0-9262-4f24-a677-c4f61362410a", "address": "fa:16:3e:03:c1:fa", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb02b2fb0-92", "ovs_interfaceid": "b02b2fb0-9262-4f24-a677-c4f61362410a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.797068] env[63372]: DEBUG nova.scheduler.client.report [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.810147] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024296, 'name': RemoveSnapshot_Task, 'duration_secs': 0.363838} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.810147] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 859.810286] env[63372]: DEBUG nova.compute.manager [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 859.811124] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40caa472-5800-4f19-8365-6bc95b966de8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.263016] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.263016] env[63372]: DEBUG nova.compute.manager [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Instance network_info: |[{"id": "b02b2fb0-9262-4f24-a677-c4f61362410a", "address": "fa:16:3e:03:c1:fa", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb02b2fb0-92", "ovs_interfaceid": "b02b2fb0-9262-4f24-a677-c4f61362410a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 860.263016] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:c1:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b02b2fb0-9262-4f24-a677-c4f61362410a', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 860.270238] env[63372]: DEBUG oslo.service.loopingcall [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.270836] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 860.271799] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63258afe-717c-4f5d-9b5a-2ba5d2e5459d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.300798] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 860.300798] env[63372]: value = "task-1024297" [ 860.300798] env[63372]: _type = "Task" [ 860.300798] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.305465] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.306185] env[63372]: DEBUG nova.compute.manager [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 860.312448] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.890s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.312790] env[63372]: DEBUG nova.objects.instance [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lazy-loading 'resources' on Instance uuid 89301344-84a9-4d13-aae7-99943d0a478e {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.314041] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024297, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.323764] env[63372]: INFO nova.compute.manager [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Shelve offloading [ 860.326446] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 860.326737] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-141cd84f-5d62-441e-9490-655536cfaaa3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.336170] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 860.336170] env[63372]: value = "task-1024298" [ 860.336170] env[63372]: _type = "Task" [ 860.336170] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.349116] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] VM already powered off {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 860.349696] env[63372]: DEBUG nova.compute.manager [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 860.350612] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf3c544-85f0-4010-9acc-4cf32a05c98a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.363136] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.363488] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.363589] env[63372]: DEBUG nova.network.neutron [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 860.557958] env[63372]: DEBUG nova.compute.manager [req-e4a62dda-07b3-4d86-93d8-a879c79cc0a7 req-476e54ec-f64b-4d8e-bcef-c9e7561990b1 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Received event network-changed-b02b2fb0-9262-4f24-a677-c4f61362410a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.558222] env[63372]: DEBUG nova.compute.manager [req-e4a62dda-07b3-4d86-93d8-a879c79cc0a7 req-476e54ec-f64b-4d8e-bcef-c9e7561990b1 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Refreshing instance network info cache due to event network-changed-b02b2fb0-9262-4f24-a677-c4f61362410a. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 860.558673] env[63372]: DEBUG oslo_concurrency.lockutils [req-e4a62dda-07b3-4d86-93d8-a879c79cc0a7 req-476e54ec-f64b-4d8e-bcef-c9e7561990b1 service nova] Acquiring lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.558673] env[63372]: DEBUG oslo_concurrency.lockutils [req-e4a62dda-07b3-4d86-93d8-a879c79cc0a7 req-476e54ec-f64b-4d8e-bcef-c9e7561990b1 service nova] Acquired lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.559025] env[63372]: DEBUG nova.network.neutron [req-e4a62dda-07b3-4d86-93d8-a879c79cc0a7 req-476e54ec-f64b-4d8e-bcef-c9e7561990b1 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Refreshing network info cache for port b02b2fb0-9262-4f24-a677-c4f61362410a {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 860.809617] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024297, 'name': CreateVM_Task, 'duration_secs': 0.409598} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.809913] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 860.810722] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.810722] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.810971] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 860.811259] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-343b86ab-ee5c-467d-be18-e95dfb4bd476 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.817118] env[63372]: DEBUG nova.compute.utils [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 860.820912] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 860.820912] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5251085c-2010-5e33-b98a-d165af9ac79c" [ 860.820912] env[63372]: _type = "Task" [ 860.820912] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.822256] env[63372]: DEBUG nova.compute.manager [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Not allocating networking since 'none' was specified. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 860.833057] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5251085c-2010-5e33-b98a-d165af9ac79c, 'name': SearchDatastore_Task, 'duration_secs': 0.010753} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.833334] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.833582] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.833773] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.833981] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.834096] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.834647] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4407e1dd-e4a8-4d38-80de-f980a392a7db {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.843736] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.844913] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 860.846566] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a538637-2259-4ee4-bd5e-78fa02f5c7cb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.859949] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 860.859949] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5272965b-b980-3c9b-4feb-cf0d22eed451" [ 860.859949] env[63372]: _type = "Task" [ 860.859949] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.870922] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5272965b-b980-3c9b-4feb-cf0d22eed451, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.112929] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7085fcf6-8a74-45d4-b284-25b6c660ebf7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.122722] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c6cafd-e684-4c2b-8eb9-b85586511d4b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.157290] env[63372]: DEBUG nova.network.neutron [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updating instance_info_cache with network_info: [{"id": "c92e8cc9-dc72-4f20-b087-1d323c502108", "address": "fa:16:3e:39:9e:be", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc92e8cc9-dc", "ovs_interfaceid": "c92e8cc9-dc72-4f20-b087-1d323c502108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.162604] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956424e6-47da-4e9b-8e23-252b3d01ea31 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.172247] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2d955e-2b55-4616-bd93-253e326d2155 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.188398] env[63372]: DEBUG nova.compute.provider_tree [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.322940] env[63372]: DEBUG nova.compute.manager [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 861.368380] env[63372]: DEBUG nova.compute.manager [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Stashing vm_state: stopped {{(pid=63372) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 861.379853] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5272965b-b980-3c9b-4feb-cf0d22eed451, 'name': SearchDatastore_Task, 'duration_secs': 0.009338} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.382113] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-335e0f1b-d44d-4873-886b-cb1f34b96340 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.390278] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 861.390278] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521725ec-3814-b882-586f-a33bc0ba68fe" [ 861.390278] env[63372]: _type = "Task" [ 861.390278] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.403276] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521725ec-3814-b882-586f-a33bc0ba68fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.566287] env[63372]: DEBUG nova.network.neutron [req-e4a62dda-07b3-4d86-93d8-a879c79cc0a7 req-476e54ec-f64b-4d8e-bcef-c9e7561990b1 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Updated VIF entry in instance network info cache for port b02b2fb0-9262-4f24-a677-c4f61362410a. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 861.570459] env[63372]: DEBUG nova.network.neutron [req-e4a62dda-07b3-4d86-93d8-a879c79cc0a7 req-476e54ec-f64b-4d8e-bcef-c9e7561990b1 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Updating instance_info_cache with network_info: [{"id": "b02b2fb0-9262-4f24-a677-c4f61362410a", "address": "fa:16:3e:03:c1:fa", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb02b2fb0-92", "ovs_interfaceid": "b02b2fb0-9262-4f24-a677-c4f61362410a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.660191] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.691458] env[63372]: DEBUG nova.scheduler.client.report [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.896596] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.902447] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521725ec-3814-b882-586f-a33bc0ba68fe, 'name': SearchDatastore_Task, 'duration_secs': 0.0101} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.902894] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.903055] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 13bde897-8446-42a2-b02d-2f5b48e6f432/13bde897-8446-42a2-b02d-2f5b48e6f432.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 861.903349] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b306ef66-41c0-476c-ad8b-ce7817c1f9cb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.910809] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 861.910809] env[63372]: value = "task-1024299" [ 861.910809] env[63372]: _type = "Task" [ 861.910809] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.920682] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024299, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.964526] env[63372]: DEBUG nova.compute.manager [req-b6a2111a-cf7b-45d7-8aa0-bc228e059a8e req-cf9a926c-0a41-4243-b8d1-591043035c78 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Received event network-vif-unplugged-c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.964771] env[63372]: DEBUG oslo_concurrency.lockutils [req-b6a2111a-cf7b-45d7-8aa0-bc228e059a8e req-cf9a926c-0a41-4243-b8d1-591043035c78 service nova] Acquiring lock "c01a5d24-eb46-4a69-993e-753880ce8e85-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.964975] env[63372]: DEBUG oslo_concurrency.lockutils [req-b6a2111a-cf7b-45d7-8aa0-bc228e059a8e req-cf9a926c-0a41-4243-b8d1-591043035c78 service nova] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.965325] env[63372]: DEBUG oslo_concurrency.lockutils [req-b6a2111a-cf7b-45d7-8aa0-bc228e059a8e req-cf9a926c-0a41-4243-b8d1-591043035c78 service nova] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.965697] env[63372]: DEBUG nova.compute.manager [req-b6a2111a-cf7b-45d7-8aa0-bc228e059a8e req-cf9a926c-0a41-4243-b8d1-591043035c78 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] No waiting events found dispatching network-vif-unplugged-c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 861.965891] env[63372]: WARNING nova.compute.manager [req-b6a2111a-cf7b-45d7-8aa0-bc228e059a8e req-cf9a926c-0a41-4243-b8d1-591043035c78 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Received unexpected event network-vif-unplugged-c92e8cc9-dc72-4f20-b087-1d323c502108 for instance with vm_state shelved and task_state shelving_offloading. [ 862.074865] env[63372]: DEBUG oslo_concurrency.lockutils [req-e4a62dda-07b3-4d86-93d8-a879c79cc0a7 req-476e54ec-f64b-4d8e-bcef-c9e7561990b1 service nova] Releasing lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.079401] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 862.080340] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a974f53-95ca-4c26-9fad-f493796ee442 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.090231] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 862.090909] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2492c586-2812-41fb-a1da-71c84e4ff3a2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.186201] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 862.186417] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 862.186643] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleting the datastore file [datastore2] c01a5d24-eb46-4a69-993e-753880ce8e85 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 862.186893] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74f9e317-fdee-4981-a5d7-c152605fac1c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.196472] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.884s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.199133] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 862.199133] env[63372]: value = "task-1024301" [ 862.199133] env[63372]: _type = "Task" [ 862.199133] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.200883] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.546s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.201215] env[63372]: DEBUG nova.objects.instance [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lazy-loading 'resources' on Instance uuid 4028d7c0-f398-4e43-9cff-5d89a14c4efa {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.221327] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024301, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.221327] env[63372]: INFO nova.scheduler.client.report [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Deleted allocations for instance 89301344-84a9-4d13-aae7-99943d0a478e [ 862.337842] env[63372]: DEBUG nova.compute.manager [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 862.372971] env[63372]: DEBUG nova.virt.hardware [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 862.373985] env[63372]: DEBUG nova.virt.hardware [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 862.374282] env[63372]: DEBUG nova.virt.hardware [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.375343] env[63372]: DEBUG nova.virt.hardware [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 862.376168] env[63372]: DEBUG nova.virt.hardware [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.376512] env[63372]: DEBUG nova.virt.hardware [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 862.376825] env[63372]: DEBUG nova.virt.hardware [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 862.377207] env[63372]: DEBUG nova.virt.hardware [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 862.377695] env[63372]: DEBUG nova.virt.hardware [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 862.378598] env[63372]: DEBUG nova.virt.hardware [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 862.379187] env[63372]: DEBUG nova.virt.hardware [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.380728] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18d39c1-7680-4665-be4d-44cb92ba4003 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.392781] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6aa3f1-30e5-467d-b397-ef8546f33974 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.408427] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Instance VIF info [] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.414300] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Creating folder: Project (4d6ead6250844639898637e53a1c5db9). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 862.415101] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-132c4e82-6e26-4d4e-bd80-11a8ec7191e5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.425826] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024299, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479336} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.426692] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 13bde897-8446-42a2-b02d-2f5b48e6f432/13bde897-8446-42a2-b02d-2f5b48e6f432.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 862.426980] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.427309] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e7149cb2-1cf8-488d-b377-644a0a991a08 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.430211] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Created folder: Project (4d6ead6250844639898637e53a1c5db9) in parent group-v227230. [ 862.430486] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Creating folder: Instances. Parent ref: group-v227385. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 862.431101] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c2977f0d-d9d4-4fbc-a183-bc9fb8a831be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.435074] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 862.435074] env[63372]: value = "task-1024303" [ 862.435074] env[63372]: _type = "Task" [ 862.435074] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.445250] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024303, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.446787] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Created folder: Instances in parent group-v227385. [ 862.447134] env[63372]: DEBUG oslo.service.loopingcall [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.447402] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 862.447701] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9672e1bf-38e1-42a3-b67d-f2fd681f83e8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.466169] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.466169] env[63372]: value = "task-1024305" [ 862.466169] env[63372]: _type = "Task" [ 862.466169] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.475018] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024305, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.717101] env[63372]: DEBUG oslo_vmware.api [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024301, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.296719} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.718123] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 862.718286] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 862.718597] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 862.741166] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ee836287-e995-4033-82c3-8c1f5b8b3a87 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "89301344-84a9-4d13-aae7-99943d0a478e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.456s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.759796] env[63372]: INFO nova.scheduler.client.report [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleted allocations for instance c01a5d24-eb46-4a69-993e-753880ce8e85 [ 862.948819] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024303, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068925} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.948819] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.949531] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87122f29-a8f6-434d-9a4a-c015710f1c73 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.973667] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 13bde897-8446-42a2-b02d-2f5b48e6f432/13bde897-8446-42a2-b02d-2f5b48e6f432.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.977314] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3a85154-60f3-4689-9eb4-8e8ad333e62b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.006960] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024305, 'name': CreateVM_Task, 'duration_secs': 0.47166} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.007475] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 863.007475] env[63372]: value = "task-1024306" [ 863.007475] env[63372]: _type = "Task" [ 863.007475] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.007705] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 863.008262] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.008468] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.008795] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 863.011883] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6993999-9d95-46a2-ac1c-de9991357e36 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.020795] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024306, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.025106] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 863.025106] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]522cb1d7-8fa0-343c-44d2-e2248368e2a6" [ 863.025106] env[63372]: _type = "Task" [ 863.025106] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.036229] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522cb1d7-8fa0-343c-44d2-e2248368e2a6, 'name': SearchDatastore_Task, 'duration_secs': 0.010977} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.036627] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.036942] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.037248] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.037430] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.037643] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.037925] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f292db2-4d8d-43d2-bac1-14b90f07200a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.050191] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.050191] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 863.050823] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d78cc568-12b8-4472-b51e-b0eba48a2a86 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.059834] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 863.059834] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52bc5805-5e5f-335e-dbef-16dabc3c7736" [ 863.059834] env[63372]: _type = "Task" [ 863.059834] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.069091] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52bc5805-5e5f-335e-dbef-16dabc3c7736, 'name': SearchDatastore_Task, 'duration_secs': 0.009976} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.071802] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae45d432-c8d1-498c-bb1e-d03897457c2a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.073581] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fb4ae13-6708-44a1-b904-a1938a172195 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.079526] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 863.079526] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52656af2-c2cd-34b3-dbd1-579116e0360d" [ 863.079526] env[63372]: _type = "Task" [ 863.079526] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.085558] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d2cd2a-57ff-43bd-9a52-ef29e237c25a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.093893] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52656af2-c2cd-34b3-dbd1-579116e0360d, 'name': SearchDatastore_Task, 'duration_secs': 0.009429} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.118567] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.118935] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] da8cf1ff-d977-4fd1-9239-043f07bed875/da8cf1ff-d977-4fd1-9239-043f07bed875.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 863.119435] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e61c32f-f58b-4757-9f5f-171c71c2aa10 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.121955] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf70dbe-c139-43c2-9a13-cd8065e34ddf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.131982] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee133796-a08c-439f-b584-832f38a49461 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.136564] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 863.136564] env[63372]: value = "task-1024307" [ 863.136564] env[63372]: _type = "Task" [ 863.136564] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.149091] env[63372]: DEBUG nova.compute.provider_tree [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.156451] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024307, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.261939] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.522388] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024306, 'name': ReconfigVM_Task, 'duration_secs': 0.328129} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.522696] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 13bde897-8446-42a2-b02d-2f5b48e6f432/13bde897-8446-42a2-b02d-2f5b48e6f432.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.523414] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-954242d8-5982-4b6c-a9c8-5c3dc8b651ed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.539052] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 863.539052] env[63372]: value = "task-1024308" [ 863.539052] env[63372]: _type = "Task" [ 863.539052] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.546660] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024308, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.653591] env[63372]: DEBUG nova.scheduler.client.report [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.660940] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024307, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501399} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.661485] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] da8cf1ff-d977-4fd1-9239-043f07bed875/da8cf1ff-d977-4fd1-9239-043f07bed875.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 863.661698] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.661960] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a43e8377-ca0a-4668-bf49-3fb13007cf9e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.670645] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 863.670645] env[63372]: value = "task-1024309" [ 863.670645] env[63372]: _type = "Task" [ 863.670645] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.691889] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024309, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.048764] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024308, 'name': Rename_Task, 'duration_secs': 0.167409} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.049294] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 864.049588] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b219ab8f-784b-4ad4-bd2c-8d1b6b0a4bec {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.057439] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 864.057439] env[63372]: value = "task-1024310" [ 864.057439] env[63372]: _type = "Task" [ 864.057439] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.067213] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024310, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.166298] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.962s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.169258] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.983s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.174027] env[63372]: INFO nova.compute.claims [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.180953] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "382fa76b-b930-459e-b84e-da1b9eb74b74" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.181219] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "382fa76b-b930-459e-b84e-da1b9eb74b74" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.188630] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024309, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069611} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.189061] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.189885] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf97125-1e66-40f5-a8e3-f7f6465d60a0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.211541] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] da8cf1ff-d977-4fd1-9239-043f07bed875/da8cf1ff-d977-4fd1-9239-043f07bed875.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.212716] env[63372]: INFO nova.scheduler.client.report [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleted allocations for instance 4028d7c0-f398-4e43-9cff-5d89a14c4efa [ 864.214397] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2b7b0c7-f8be-4a07-b125-09d58d671763 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.237625] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 864.237625] env[63372]: value = "task-1024311" [ 864.237625] env[63372]: _type = "Task" [ 864.237625] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.247447] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024311, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.402345] env[63372]: DEBUG nova.compute.manager [req-a4703e1b-86fa-4391-bddb-364a8f497c55 req-a3b63011-19f1-4c63-8ad6-53b0c28cb277 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Received event network-changed-c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.402345] env[63372]: DEBUG nova.compute.manager [req-a4703e1b-86fa-4391-bddb-364a8f497c55 req-a3b63011-19f1-4c63-8ad6-53b0c28cb277 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Refreshing instance network info cache due to event network-changed-c92e8cc9-dc72-4f20-b087-1d323c502108. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 864.402345] env[63372]: DEBUG oslo_concurrency.lockutils [req-a4703e1b-86fa-4391-bddb-364a8f497c55 req-a3b63011-19f1-4c63-8ad6-53b0c28cb277 service nova] Acquiring lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.402345] env[63372]: DEBUG oslo_concurrency.lockutils [req-a4703e1b-86fa-4391-bddb-364a8f497c55 req-a3b63011-19f1-4c63-8ad6-53b0c28cb277 service nova] Acquired lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.402345] env[63372]: DEBUG nova.network.neutron [req-a4703e1b-86fa-4391-bddb-364a8f497c55 req-a3b63011-19f1-4c63-8ad6-53b0c28cb277 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Refreshing network info cache for port c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 864.571251] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024310, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.683456] env[63372]: DEBUG nova.compute.manager [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 864.737267] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6cf34705-2b06-4a16-9802-59b03549630e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "4028d7c0-f398-4e43-9cff-5d89a14c4efa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.925s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.751033] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.069924] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024310, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.211426] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.258281] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024311, 'name': ReconfigVM_Task, 'duration_secs': 0.725271} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.264734] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Reconfigured VM instance instance-0000004b to attach disk [datastore1] da8cf1ff-d977-4fd1-9239-043f07bed875/da8cf1ff-d977-4fd1-9239-043f07bed875.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.265942] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-871a30c4-c9e7-4551-b02a-f71e57c6d36c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.276267] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 865.276267] env[63372]: value = "task-1024312" [ 865.276267] env[63372]: _type = "Task" [ 865.276267] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.297503] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024312, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.453172] env[63372]: DEBUG nova.network.neutron [req-a4703e1b-86fa-4391-bddb-364a8f497c55 req-a3b63011-19f1-4c63-8ad6-53b0c28cb277 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updated VIF entry in instance network info cache for port c92e8cc9-dc72-4f20-b087-1d323c502108. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 865.453879] env[63372]: DEBUG nova.network.neutron [req-a4703e1b-86fa-4391-bddb-364a8f497c55 req-a3b63011-19f1-4c63-8ad6-53b0c28cb277 service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updating instance_info_cache with network_info: [{"id": "c92e8cc9-dc72-4f20-b087-1d323c502108", "address": "fa:16:3e:39:9e:be", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapc92e8cc9-dc", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.571251] env[63372]: DEBUG oslo_vmware.api [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024310, 'name': PowerOnVM_Task, 'duration_secs': 1.034156} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.574315] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 865.575324] env[63372]: INFO nova.compute.manager [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Took 7.50 seconds to spawn the instance on the hypervisor. [ 865.575324] env[63372]: DEBUG nova.compute.manager [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.575681] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d279f174-9f6f-49b3-bb83-ce4559e48b15 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.589854] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1495f691-b5ac-4a30-bf3b-bbcfcb86b42f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.599514] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-291a1972-9f7b-4db7-b483-264b96f68f1b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.607766] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "c01a5d24-eb46-4a69-993e-753880ce8e85" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.636293] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c2be3e1-0e1e-4f99-bc39-b69bdef56fa0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.645490] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e70af7-632c-440c-a9c4-42cce5f4ad3c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.662292] env[63372]: DEBUG nova.compute.provider_tree [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.788834] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024312, 'name': Rename_Task, 'duration_secs': 0.278336} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.789207] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 865.790509] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d878001e-2a3d-43e1-b9ca-73aa14069613 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.799979] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 865.799979] env[63372]: value = "task-1024313" [ 865.799979] env[63372]: _type = "Task" [ 865.799979] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.810187] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024313, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.961931] env[63372]: DEBUG oslo_concurrency.lockutils [req-a4703e1b-86fa-4391-bddb-364a8f497c55 req-a3b63011-19f1-4c63-8ad6-53b0c28cb277 service nova] Releasing lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.990895] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "75560852-8c6d-4220-b251-81e8721a4b7b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.992048] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "75560852-8c6d-4220-b251-81e8721a4b7b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.103194] env[63372]: INFO nova.compute.manager [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Took 30.40 seconds to build instance. [ 866.165459] env[63372]: DEBUG nova.scheduler.client.report [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.190240] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "ba1daa0a-dcf8-4586-9964-a962fcc94196" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.190632] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "ba1daa0a-dcf8-4586-9964-a962fcc94196" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.315317] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024313, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.494825] env[63372]: DEBUG nova.compute.manager [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 866.608481] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2145699-fb7a-43bb-92de-4d1d8fe27466 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "13bde897-8446-42a2-b02d-2f5b48e6f432" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.921s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.672081] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.672647] env[63372]: DEBUG nova.compute.manager [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 866.680528] env[63372]: DEBUG oslo_concurrency.lockutils [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.081s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.680777] env[63372]: DEBUG nova.objects.instance [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lazy-loading 'resources' on Instance uuid de1b38a4-c7f2-420c-a050-7311976e4ca8 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.692767] env[63372]: DEBUG nova.compute.manager [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 866.818758] env[63372]: DEBUG oslo_vmware.api [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024313, 'name': PowerOnVM_Task, 'duration_secs': 0.772783} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.820675] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 866.820675] env[63372]: INFO nova.compute.manager [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Took 4.48 seconds to spawn the instance on the hypervisor. [ 866.820675] env[63372]: DEBUG nova.compute.manager [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.821298] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4db85a5-e19b-40a8-ab5d-5990bdf08625 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.018048] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.192226] env[63372]: DEBUG nova.compute.utils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 867.195266] env[63372]: DEBUG nova.compute.manager [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 867.195597] env[63372]: DEBUG nova.network.neutron [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 867.203829] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "d568b727-7836-4b1d-9f27-f159227e46f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.203829] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "d568b727-7836-4b1d-9f27-f159227e46f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.224741] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.303768] env[63372]: DEBUG nova.policy [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '75c76e021bd94121b4bb416bb7845bda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '624e1cee468d4a538c70adf2e058869c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 867.349616] env[63372]: INFO nova.compute.manager [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Took 29.33 seconds to build instance. [ 867.539890] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "326a9155-1c00-4c18-a9a7-6124e709c20b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.540106] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "326a9155-1c00-4c18-a9a7-6124e709c20b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.626059] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ad1b15-530e-4544-9ff3-5b5d7af1c63f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.637418] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff8a75f-cbae-4022-ac4e-7e08abb50283 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.645036] env[63372]: DEBUG nova.compute.manager [req-4cdcdac7-2034-4054-9d6a-ebc06fa24a39 req-d8c2eaaa-c691-4892-b903-02809b021a97 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Received event network-changed-b02b2fb0-9262-4f24-a677-c4f61362410a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.646070] env[63372]: DEBUG nova.compute.manager [req-4cdcdac7-2034-4054-9d6a-ebc06fa24a39 req-d8c2eaaa-c691-4892-b903-02809b021a97 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Refreshing instance network info cache due to event network-changed-b02b2fb0-9262-4f24-a677-c4f61362410a. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 867.647783] env[63372]: DEBUG oslo_concurrency.lockutils [req-4cdcdac7-2034-4054-9d6a-ebc06fa24a39 req-d8c2eaaa-c691-4892-b903-02809b021a97 service nova] Acquiring lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.647783] env[63372]: DEBUG oslo_concurrency.lockutils [req-4cdcdac7-2034-4054-9d6a-ebc06fa24a39 req-d8c2eaaa-c691-4892-b903-02809b021a97 service nova] Acquired lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.647985] env[63372]: DEBUG nova.network.neutron [req-4cdcdac7-2034-4054-9d6a-ebc06fa24a39 req-d8c2eaaa-c691-4892-b903-02809b021a97 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Refreshing network info cache for port b02b2fb0-9262-4f24-a677-c4f61362410a {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 867.679957] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6624818-dacf-48c9-bbc7-bec9d0e01e91 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.690930] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682f12e1-2951-4884-a2f0-57502002ab93 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.696321] env[63372]: DEBUG nova.compute.manager [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 867.712332] env[63372]: DEBUG nova.compute.manager [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 867.715023] env[63372]: DEBUG nova.compute.provider_tree [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.758137] env[63372]: INFO nova.compute.manager [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Rebuilding instance [ 867.810287] env[63372]: DEBUG nova.network.neutron [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Successfully created port: 75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 867.816151] env[63372]: DEBUG nova.compute.manager [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.817143] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b368ed9-8662-4067-9f48-03b39e62d0ae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.853867] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d69f130c-80e9-4418-9a22-e1b555526fe2 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Lock "da8cf1ff-d977-4fd1-9239-043f07bed875" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.839s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.043553] env[63372]: DEBUG nova.compute.manager [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 868.223095] env[63372]: DEBUG nova.scheduler.client.report [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.249063] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.331157] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 868.331482] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d5b1fcc-726d-4a69-9ad8-786b03844674 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.341289] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 868.341289] env[63372]: value = "task-1024314" [ 868.341289] env[63372]: _type = "Task" [ 868.341289] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.350618] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024314, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.464014] env[63372]: DEBUG nova.network.neutron [req-4cdcdac7-2034-4054-9d6a-ebc06fa24a39 req-d8c2eaaa-c691-4892-b903-02809b021a97 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Updated VIF entry in instance network info cache for port b02b2fb0-9262-4f24-a677-c4f61362410a. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 868.464664] env[63372]: DEBUG nova.network.neutron [req-4cdcdac7-2034-4054-9d6a-ebc06fa24a39 req-d8c2eaaa-c691-4892-b903-02809b021a97 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Updating instance_info_cache with network_info: [{"id": "b02b2fb0-9262-4f24-a677-c4f61362410a", "address": "fa:16:3e:03:c1:fa", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb02b2fb0-92", "ovs_interfaceid": "b02b2fb0-9262-4f24-a677-c4f61362410a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.565221] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.710750] env[63372]: DEBUG nova.compute.manager [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 868.732638] env[63372]: DEBUG oslo_concurrency.lockutils [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.052s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.734735] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.560s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.735556] env[63372]: DEBUG nova.objects.instance [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lazy-loading 'resources' on Instance uuid a3d5b93e-1b76-48e7-bb3e-2423e7c4119a {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.738504] env[63372]: DEBUG nova.virt.hardware [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 868.738504] env[63372]: DEBUG nova.virt.hardware [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 868.738678] env[63372]: DEBUG nova.virt.hardware [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.738868] env[63372]: DEBUG nova.virt.hardware [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 868.739031] env[63372]: DEBUG nova.virt.hardware [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.739188] env[63372]: DEBUG nova.virt.hardware [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 868.739395] env[63372]: DEBUG nova.virt.hardware [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 868.739575] env[63372]: DEBUG nova.virt.hardware [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 868.739757] env[63372]: DEBUG nova.virt.hardware [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 868.739952] env[63372]: DEBUG nova.virt.hardware [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 868.740104] env[63372]: DEBUG nova.virt.hardware [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.741433] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3a0749-467f-43e7-9faa-f33eb516b6df {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.750966] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e4b137-5823-42c0-b87e-06b2a8365c9a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.755606] env[63372]: INFO nova.scheduler.client.report [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Deleted allocations for instance de1b38a4-c7f2-420c-a050-7311976e4ca8 [ 868.850972] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024314, 'name': PowerOffVM_Task, 'duration_secs': 0.135502} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.852076] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 868.852076] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 868.852433] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c4244a-ade3-47f4-bbb9-03bab72bbba2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.858820] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 868.859039] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22bb798a-6145-4b90-b2fc-4fe08f2e35f9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.883858] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 868.884160] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 868.884393] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Deleting the datastore file [datastore1] da8cf1ff-d977-4fd1-9239-043f07bed875 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 868.884660] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec99ff4a-6e5f-4efb-98f7-3e4091347165 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.891230] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 868.891230] env[63372]: value = "task-1024316" [ 868.891230] env[63372]: _type = "Task" [ 868.891230] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.899345] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.967738] env[63372]: DEBUG oslo_concurrency.lockutils [req-4cdcdac7-2034-4054-9d6a-ebc06fa24a39 req-d8c2eaaa-c691-4892-b903-02809b021a97 service nova] Releasing lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.267825] env[63372]: DEBUG oslo_concurrency.lockutils [None req-57bd1a07-8723-4c03-995e-73ba7c935d8c tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "de1b38a4-c7f2-420c-a050-7311976e4ca8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.679s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.404155] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024316, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.086079} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.407021] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.407021] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 869.407021] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 869.540111] env[63372]: DEBUG nova.network.neutron [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Successfully updated port: 75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 869.558719] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6074e498-b059-4a7a-bd93-d4dfc31e0ae7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.572380] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53274c6b-a532-466c-9c1b-38e3b59da56c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.604096] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e6a2f1-ec59-49d6-a4bb-2111b07a4d32 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.607071] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "99f901a6-9bb3-4403-af0c-c8900f655cb3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.607166] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "99f901a6-9bb3-4403-af0c-c8900f655cb3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.607366] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "99f901a6-9bb3-4403-af0c-c8900f655cb3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.607552] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "99f901a6-9bb3-4403-af0c-c8900f655cb3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.607780] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "99f901a6-9bb3-4403-af0c-c8900f655cb3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.610227] env[63372]: INFO nova.compute.manager [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Terminating instance [ 869.612722] env[63372]: DEBUG nova.compute.manager [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 869.612923] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 869.615510] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59fbeb1e-807e-4f1b-94cc-54ca90d0b9a4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.619251] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f97489d2-635c-4ca0-9a9d-d5a7795c48cd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.628163] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 869.636708] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6f3b9a5-213e-4b33-9ae6-d0dfc697f3cb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.638720] env[63372]: DEBUG nova.compute.provider_tree [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.646299] env[63372]: DEBUG oslo_vmware.api [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 869.646299] env[63372]: value = "task-1024317" [ 869.646299] env[63372]: _type = "Task" [ 869.646299] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.657346] env[63372]: DEBUG oslo_vmware.api [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024317, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.706263] env[63372]: DEBUG nova.compute.manager [req-0c549191-b86a-4dad-8d97-8e2aba1bad52 req-137e5bff-ddd3-499e-80ae-2b433cbd54ad service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Received event network-vif-plugged-75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.706676] env[63372]: DEBUG oslo_concurrency.lockutils [req-0c549191-b86a-4dad-8d97-8e2aba1bad52 req-137e5bff-ddd3-499e-80ae-2b433cbd54ad service nova] Acquiring lock "1f52f9ee-2af3-4321-8ec6-5d5e841ed880-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.706811] env[63372]: DEBUG oslo_concurrency.lockutils [req-0c549191-b86a-4dad-8d97-8e2aba1bad52 req-137e5bff-ddd3-499e-80ae-2b433cbd54ad service nova] Lock "1f52f9ee-2af3-4321-8ec6-5d5e841ed880-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.706985] env[63372]: DEBUG oslo_concurrency.lockutils [req-0c549191-b86a-4dad-8d97-8e2aba1bad52 req-137e5bff-ddd3-499e-80ae-2b433cbd54ad service nova] Lock "1f52f9ee-2af3-4321-8ec6-5d5e841ed880-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.707169] env[63372]: DEBUG nova.compute.manager [req-0c549191-b86a-4dad-8d97-8e2aba1bad52 req-137e5bff-ddd3-499e-80ae-2b433cbd54ad service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] No waiting events found dispatching network-vif-plugged-75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 869.707335] env[63372]: WARNING nova.compute.manager [req-0c549191-b86a-4dad-8d97-8e2aba1bad52 req-137e5bff-ddd3-499e-80ae-2b433cbd54ad service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Received unexpected event network-vif-plugged-75cfde99-66f3-4e6e-8755-061b37055d28 for instance with vm_state building and task_state spawning. [ 869.707498] env[63372]: DEBUG nova.compute.manager [req-0c549191-b86a-4dad-8d97-8e2aba1bad52 req-137e5bff-ddd3-499e-80ae-2b433cbd54ad service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Received event network-changed-75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.707655] env[63372]: DEBUG nova.compute.manager [req-0c549191-b86a-4dad-8d97-8e2aba1bad52 req-137e5bff-ddd3-499e-80ae-2b433cbd54ad service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Refreshing instance network info cache due to event network-changed-75cfde99-66f3-4e6e-8755-061b37055d28. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 869.707834] env[63372]: DEBUG oslo_concurrency.lockutils [req-0c549191-b86a-4dad-8d97-8e2aba1bad52 req-137e5bff-ddd3-499e-80ae-2b433cbd54ad service nova] Acquiring lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.707969] env[63372]: DEBUG oslo_concurrency.lockutils [req-0c549191-b86a-4dad-8d97-8e2aba1bad52 req-137e5bff-ddd3-499e-80ae-2b433cbd54ad service nova] Acquired lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.708135] env[63372]: DEBUG nova.network.neutron [req-0c549191-b86a-4dad-8d97-8e2aba1bad52 req-137e5bff-ddd3-499e-80ae-2b433cbd54ad service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Refreshing network info cache for port 75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 870.046799] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquiring lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.141730] env[63372]: DEBUG nova.scheduler.client.report [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 870.157732] env[63372]: DEBUG oslo_vmware.api [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024317, 'name': PowerOffVM_Task, 'duration_secs': 0.280697} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.158630] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 870.158820] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 870.159080] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-259f661c-f718-4a66-aec1-a3bc58b2f563 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.240468] env[63372]: DEBUG nova.network.neutron [req-0c549191-b86a-4dad-8d97-8e2aba1bad52 req-137e5bff-ddd3-499e-80ae-2b433cbd54ad service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 870.250380] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 870.251310] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 870.251310] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Deleting the datastore file [datastore1] 99f901a6-9bb3-4403-af0c-c8900f655cb3 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 870.251310] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e1cbbfc9-d4c4-4d55-9d48-7baba38bb9b0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.258607] env[63372]: DEBUG oslo_vmware.api [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for the task: (returnval){ [ 870.258607] env[63372]: value = "task-1024319" [ 870.258607] env[63372]: _type = "Task" [ 870.258607] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.266925] env[63372]: DEBUG oslo_vmware.api [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024319, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.327375] env[63372]: DEBUG nova.network.neutron [req-0c549191-b86a-4dad-8d97-8e2aba1bad52 req-137e5bff-ddd3-499e-80ae-2b433cbd54ad service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.444548] env[63372]: DEBUG nova.virt.hardware [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.444900] env[63372]: DEBUG nova.virt.hardware [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.445629] env[63372]: DEBUG nova.virt.hardware [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.445942] env[63372]: DEBUG nova.virt.hardware [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.446176] env[63372]: DEBUG nova.virt.hardware [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.446388] env[63372]: DEBUG nova.virt.hardware [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.446690] env[63372]: DEBUG nova.virt.hardware [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.446931] env[63372]: DEBUG nova.virt.hardware [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.447113] env[63372]: DEBUG nova.virt.hardware [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.447343] env[63372]: DEBUG nova.virt.hardware [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.447589] env[63372]: DEBUG nova.virt.hardware [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.449207] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e596a0-bcf8-4875-850b-3b3a2e2b402f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.457835] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e52296-eb07-47d3-bef9-550c715ad924 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.475279] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Instance VIF info [] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.481284] env[63372]: DEBUG oslo.service.loopingcall [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.481579] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 870.481981] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-50309370-ec9b-4cc3-b3b8-8819a9e6ac57 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.499645] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.499645] env[63372]: value = "task-1024320" [ 870.499645] env[63372]: _type = "Task" [ 870.499645] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.509488] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024320, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.646510] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.912s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.649934] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.919s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.652038] env[63372]: INFO nova.compute.claims [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 870.672172] env[63372]: INFO nova.scheduler.client.report [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Deleted allocations for instance a3d5b93e-1b76-48e7-bb3e-2423e7c4119a [ 870.773328] env[63372]: DEBUG oslo_vmware.api [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Task: {'id': task-1024319, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167315} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.773328] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.774709] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 870.774709] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 870.774709] env[63372]: INFO nova.compute.manager [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Took 1.16 seconds to destroy the instance on the hypervisor. [ 870.775819] env[63372]: DEBUG oslo.service.loopingcall [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.775819] env[63372]: DEBUG nova.compute.manager [-] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 870.775819] env[63372]: DEBUG nova.network.neutron [-] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 870.830736] env[63372]: DEBUG oslo_concurrency.lockutils [req-0c549191-b86a-4dad-8d97-8e2aba1bad52 req-137e5bff-ddd3-499e-80ae-2b433cbd54ad service nova] Releasing lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.831160] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquired lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.831323] env[63372]: DEBUG nova.network.neutron [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 871.010477] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024320, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.180336] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89aa241a-15b3-4638-93e7-e9911655b009 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "a3d5b93e-1b76-48e7-bb3e-2423e7c4119a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.091s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.382832] env[63372]: DEBUG nova.network.neutron [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 871.519133] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024320, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.526817] env[63372]: DEBUG nova.network.neutron [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Updating instance_info_cache with network_info: [{"id": "75cfde99-66f3-4e6e-8755-061b37055d28", "address": "fa:16:3e:b9:fd:d5", "network": {"id": "00a725f6-e5e4-463e-a8ff-3a0953fd3186", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2027272115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "624e1cee468d4a538c70adf2e058869c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9950022e-0005-4da9-b1c4-6c9fec5ea180", "external-id": "nsx-vlan-transportzone-448", "segmentation_id": 448, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75cfde99-66", "ovs_interfaceid": "75cfde99-66f3-4e6e-8755-061b37055d28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.601291] env[63372]: DEBUG nova.network.neutron [-] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.743287] env[63372]: DEBUG nova.compute.manager [req-4714cd91-bc58-40b0-82fd-6ca478f5143a req-a711ad9f-3495-4f15-80e3-6d58a5501ec4 service nova] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Received event network-vif-deleted-0e3060b6-6d21-4a2d-8e8a-ae48e64ee869 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.911228] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7182a7-3bab-4a47-a1df-2d57e40561f1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.919615] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03fc7113-1f45-48d9-aea2-8b01f5a0ca2a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.952832] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc178f51-0387-40bf-a0e0-5a88aa2621c2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.961240] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecabc01-51b5-4b2a-94e4-043fb4c0ac6a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.975499] env[63372]: DEBUG nova.compute.provider_tree [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 872.011787] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024320, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.029902] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Releasing lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.030230] env[63372]: DEBUG nova.compute.manager [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Instance network_info: |[{"id": "75cfde99-66f3-4e6e-8755-061b37055d28", "address": "fa:16:3e:b9:fd:d5", "network": {"id": "00a725f6-e5e4-463e-a8ff-3a0953fd3186", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2027272115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "624e1cee468d4a538c70adf2e058869c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9950022e-0005-4da9-b1c4-6c9fec5ea180", "external-id": "nsx-vlan-transportzone-448", "segmentation_id": 448, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75cfde99-66", "ovs_interfaceid": "75cfde99-66f3-4e6e-8755-061b37055d28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 872.030638] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:fd:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9950022e-0005-4da9-b1c4-6c9fec5ea180', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '75cfde99-66f3-4e6e-8755-061b37055d28', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.038192] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Creating folder: Project (624e1cee468d4a538c70adf2e058869c). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 872.038453] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bc29460b-c5f8-46b0-97c1-1ac4fb32a763 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.050654] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Created folder: Project (624e1cee468d4a538c70adf2e058869c) in parent group-v227230. [ 872.050879] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Creating folder: Instances. Parent ref: group-v227389. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 872.051143] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5d7d11a7-f126-47b9-8f19-e3cf4a9508c5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.061273] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Created folder: Instances in parent group-v227389. [ 872.061518] env[63372]: DEBUG oslo.service.loopingcall [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.062504] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 872.062504] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e4296859-4647-4a32-9386-c7dc215fa9c4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.082206] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.082206] env[63372]: value = "task-1024324" [ 872.082206] env[63372]: _type = "Task" [ 872.082206] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.090227] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024324, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.103908] env[63372]: INFO nova.compute.manager [-] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Took 1.33 seconds to deallocate network for instance. [ 872.495454] env[63372]: ERROR nova.scheduler.client.report [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [req-ff031fda-88c2-460d-85c0-e94058379120] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ff031fda-88c2-460d-85c0-e94058379120"}]} [ 872.512287] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024320, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.513315] env[63372]: DEBUG nova.scheduler.client.report [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Refreshing inventories for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 872.527213] env[63372]: DEBUG nova.scheduler.client.report [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Updating ProviderTree inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 872.527516] env[63372]: DEBUG nova.compute.provider_tree [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 872.540075] env[63372]: DEBUG nova.scheduler.client.report [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Refreshing aggregate associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, aggregates: None {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 872.560610] env[63372]: DEBUG nova.scheduler.client.report [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Refreshing trait associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 872.594393] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024324, 'name': CreateVM_Task, 'duration_secs': 0.3521} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.594580] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 872.595286] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.595461] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.595787] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 872.596056] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c3ce114-f330-41a7-927e-39190640831f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.601294] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 872.601294] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5255e0be-75dc-7358-5c4f-fc481dd383d9" [ 872.601294] env[63372]: _type = "Task" [ 872.601294] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.612228] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.612515] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5255e0be-75dc-7358-5c4f-fc481dd383d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.833180] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a3b94f-2a50-4c63-bda9-3aa3656e1aad {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.842156] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-822b60d8-8e40-45a3-b329-62b23341a505 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.876415] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bada7dd-2343-401d-b74d-b275bdaa2e20 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.886368] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a6570c-b121-4901-a9d3-9c1aa65b9877 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.901539] env[63372]: DEBUG nova.compute.provider_tree [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 873.013525] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024320, 'name': CreateVM_Task, 'duration_secs': 2.094315} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.013719] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 873.014222] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.014398] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.014715] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.014980] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad34ae50-f3ad-4c96-85cb-52e1d7d569d3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.020574] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 873.020574] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52adea2c-6434-bbeb-80f0-020e0830ee3f" [ 873.020574] env[63372]: _type = "Task" [ 873.020574] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.029891] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52adea2c-6434-bbeb-80f0-020e0830ee3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.114528] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5255e0be-75dc-7358-5c4f-fc481dd383d9, 'name': SearchDatastore_Task, 'duration_secs': 0.011005} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.114844] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.115144] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.115385] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.115535] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.115730] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.115969] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5753dea9-19b6-4ddf-9958-09fd3f09c69c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.124930] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.125126] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 873.125828] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-974ba805-d4a9-4a73-85b2-cbb1553b46b9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.131181] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 873.131181] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5240c837-d637-e343-94af-3c29f1141d3d" [ 873.131181] env[63372]: _type = "Task" [ 873.131181] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.139203] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5240c837-d637-e343-94af-3c29f1141d3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.433959] env[63372]: DEBUG nova.scheduler.client.report [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 105 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 873.434261] env[63372]: DEBUG nova.compute.provider_tree [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 105 to 106 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 873.434461] env[63372]: DEBUG nova.compute.provider_tree [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 873.531836] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52adea2c-6434-bbeb-80f0-020e0830ee3f, 'name': SearchDatastore_Task, 'duration_secs': 0.010493} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.532166] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.532400] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.532632] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.532782] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.532963] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.533242] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7aa629e6-2db7-4a02-870e-ed9552f1fc83 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.541574] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.541755] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 873.542458] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac44eed6-4830-474b-a1b3-ff3997cb26df {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.547442] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 873.547442] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52647abe-dd5c-27d2-a39b-5a37675a3c74" [ 873.547442] env[63372]: _type = "Task" [ 873.547442] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.554724] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52647abe-dd5c-27d2-a39b-5a37675a3c74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.641947] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5240c837-d637-e343-94af-3c29f1141d3d, 'name': SearchDatastore_Task, 'duration_secs': 0.008399} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.642750] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-432e56f0-fdae-4dbf-b97e-30ea9a13023c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.648988] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 873.648988] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5276b02e-ba8f-57d6-b713-9b69c7b724dd" [ 873.648988] env[63372]: _type = "Task" [ 873.648988] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.658683] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5276b02e-ba8f-57d6-b713-9b69c7b724dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.940129] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.290s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.940491] env[63372]: DEBUG nova.compute.manager [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 873.943377] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.825s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.943592] env[63372]: DEBUG nova.objects.instance [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lazy-loading 'resources' on Instance uuid ae6f3f42-7213-4ab1-b74c-1a557df6748b {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.060263] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52647abe-dd5c-27d2-a39b-5a37675a3c74, 'name': SearchDatastore_Task, 'duration_secs': 0.011643} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.061117] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a52e6fe-75aa-4f02-b14a-c7dba46b279a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.068376] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 874.068376] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]527ea62f-6115-eca4-38b4-bebe111be84d" [ 874.068376] env[63372]: _type = "Task" [ 874.068376] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.079829] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527ea62f-6115-eca4-38b4-bebe111be84d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.159581] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5276b02e-ba8f-57d6-b713-9b69c7b724dd, 'name': SearchDatastore_Task, 'duration_secs': 0.010055} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.159944] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.160233] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 1f52f9ee-2af3-4321-8ec6-5d5e841ed880/1f52f9ee-2af3-4321-8ec6-5d5e841ed880.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 874.160499] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6367526b-40b7-4169-a07b-48b2e166c50c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.168871] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 874.168871] env[63372]: value = "task-1024325" [ 874.168871] env[63372]: _type = "Task" [ 874.168871] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.176719] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024325, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.381290] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "ac90a156-be00-4f62-a76e-e08914531167" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.381497] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "ac90a156-be00-4f62-a76e-e08914531167" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.381726] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "ac90a156-be00-4f62-a76e-e08914531167-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.381919] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "ac90a156-be00-4f62-a76e-e08914531167-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.382113] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "ac90a156-be00-4f62-a76e-e08914531167-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.384540] env[63372]: INFO nova.compute.manager [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Terminating instance [ 874.386846] env[63372]: DEBUG nova.compute.manager [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 874.387078] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 874.388013] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43bff138-c8b3-4f4c-9c1a-581a8030d908 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.398030] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 874.398209] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-876c2b13-2b1d-4c94-8770-a8e29253a148 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.406646] env[63372]: DEBUG oslo_vmware.api [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 874.406646] env[63372]: value = "task-1024326" [ 874.406646] env[63372]: _type = "Task" [ 874.406646] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.417178] env[63372]: DEBUG oslo_vmware.api [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024326, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.447430] env[63372]: DEBUG nova.compute.utils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 874.449455] env[63372]: DEBUG nova.compute.manager [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 874.449709] env[63372]: DEBUG nova.network.neutron [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 874.582369] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527ea62f-6115-eca4-38b4-bebe111be84d, 'name': SearchDatastore_Task, 'duration_secs': 0.010104} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.582369] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.582369] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] da8cf1ff-d977-4fd1-9239-043f07bed875/da8cf1ff-d977-4fd1-9239-043f07bed875.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 874.582898] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-307bd58f-43f8-467d-b3bd-62ac0f4b9182 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.590918] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 874.590918] env[63372]: value = "task-1024327" [ 874.590918] env[63372]: _type = "Task" [ 874.590918] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.603991] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024327, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.683263] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024325, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486717} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.686429] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 1f52f9ee-2af3-4321-8ec6-5d5e841ed880/1f52f9ee-2af3-4321-8ec6-5d5e841ed880.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 874.686667] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 874.687144] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e5e4792-39ef-41ff-8f19-d4958d07e8f5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.695207] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 874.695207] env[63372]: value = "task-1024328" [ 874.695207] env[63372]: _type = "Task" [ 874.695207] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.707781] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024328, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.754849] env[63372]: DEBUG nova.policy [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f8e364e3aed40248cc0d11741f76154', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b4548b52cd704cc0b054e2f7d1562b9a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 874.777921] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0726ae79-a3c6-4689-a5d2-4a3ada8f2994 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.786103] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6059281b-2b1e-4bb8-9880-69bc13d292c2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.818124] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe25cd81-ba0e-4c68-bced-55aa7114cc14 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.828262] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8885bce0-4ca3-48ef-825a-feee3c0c6f73 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.843083] env[63372]: DEBUG nova.compute.provider_tree [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.918793] env[63372]: DEBUG oslo_vmware.api [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024326, 'name': PowerOffVM_Task, 'duration_secs': 0.311389} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.918896] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 874.919084] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 874.919366] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-71f9e8a8-ff0c-4fe3-8504-2de9e587cfa5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.957258] env[63372]: DEBUG nova.compute.manager [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 874.997008] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 874.997196] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 874.997350] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Deleting the datastore file [datastore2] ac90a156-be00-4f62-a76e-e08914531167 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.997775] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d6769bba-f2e5-4a32-a4b3-cff03752c487 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.005664] env[63372]: DEBUG oslo_vmware.api [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for the task: (returnval){ [ 875.005664] env[63372]: value = "task-1024330" [ 875.005664] env[63372]: _type = "Task" [ 875.005664] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.016060] env[63372]: DEBUG oslo_vmware.api [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024330, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.103846] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024327, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.147407] env[63372]: DEBUG nova.network.neutron [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Successfully created port: 617ca099-5dd5-46c3-afc2-f80f4dbe242f {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 875.211304] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024328, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069456} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.211651] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 875.212488] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca04a19c-f424-420b-ae8b-7a4cf847a580 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.239509] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 1f52f9ee-2af3-4321-8ec6-5d5e841ed880/1f52f9ee-2af3-4321-8ec6-5d5e841ed880.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.239509] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d608db4-7c91-41ad-9b55-952cc79127d8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.261985] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 875.261985] env[63372]: value = "task-1024331" [ 875.261985] env[63372]: _type = "Task" [ 875.261985] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.273014] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024331, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.346275] env[63372]: DEBUG nova.scheduler.client.report [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 875.517987] env[63372]: DEBUG oslo_vmware.api [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Task: {'id': task-1024330, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.219894} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.518317] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.518546] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 875.518767] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 875.518939] env[63372]: INFO nova.compute.manager [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] [instance: ac90a156-be00-4f62-a76e-e08914531167] Took 1.13 seconds to destroy the instance on the hypervisor. [ 875.523361] env[63372]: DEBUG oslo.service.loopingcall [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.523831] env[63372]: DEBUG nova.compute.manager [-] [instance: ac90a156-be00-4f62-a76e-e08914531167] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 875.523937] env[63372]: DEBUG nova.network.neutron [-] [instance: ac90a156-be00-4f62-a76e-e08914531167] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 875.601920] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024327, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.537532} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.602233] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] da8cf1ff-d977-4fd1-9239-043f07bed875/da8cf1ff-d977-4fd1-9239-043f07bed875.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 875.602422] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.603213] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a35a8ac7-b03b-4015-8ac9-f9734df60fda {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.612526] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 875.612526] env[63372]: value = "task-1024332" [ 875.612526] env[63372]: _type = "Task" [ 875.612526] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.624493] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024332, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.778317] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024331, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.851847] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.908s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.854729] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.383s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.855125] env[63372]: DEBUG nova.objects.instance [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Lazy-loading 'resources' on Instance uuid 07dab964-a741-4723-98a3-0cfbe4eff975 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.883623] env[63372]: INFO nova.scheduler.client.report [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Deleted allocations for instance ae6f3f42-7213-4ab1-b74c-1a557df6748b [ 875.973556] env[63372]: DEBUG nova.compute.manager [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 876.005694] env[63372]: DEBUG nova.virt.hardware [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.005994] env[63372]: DEBUG nova.virt.hardware [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.006329] env[63372]: DEBUG nova.virt.hardware [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.006329] env[63372]: DEBUG nova.virt.hardware [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.006434] env[63372]: DEBUG nova.virt.hardware [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.006561] env[63372]: DEBUG nova.virt.hardware [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.006757] env[63372]: DEBUG nova.virt.hardware [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.007714] env[63372]: DEBUG nova.virt.hardware [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.007714] env[63372]: DEBUG nova.virt.hardware [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.007714] env[63372]: DEBUG nova.virt.hardware [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.007714] env[63372]: DEBUG nova.virt.hardware [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.008337] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8fd5d14-5c3f-4df9-9168-04026b7ed5d5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.017628] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f10bdd4-5fe6-4bf0-8dcf-b52defd5e49b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.035033] env[63372]: DEBUG nova.compute.manager [req-4a7ddf2c-a694-4d75-80fe-127f39536504 req-be4e1316-3877-457b-9212-30443aa9be17 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Received event network-vif-deleted-40415ba1-4a9b-45d8-90ff-d95ac4e3980f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.035264] env[63372]: INFO nova.compute.manager [req-4a7ddf2c-a694-4d75-80fe-127f39536504 req-be4e1316-3877-457b-9212-30443aa9be17 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Neutron deleted interface 40415ba1-4a9b-45d8-90ff-d95ac4e3980f; detaching it from the instance and deleting it from the info cache [ 876.035438] env[63372]: DEBUG nova.network.neutron [req-4a7ddf2c-a694-4d75-80fe-127f39536504 req-be4e1316-3877-457b-9212-30443aa9be17 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.124291] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024332, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.24708} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.124592] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.125376] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b88a18-9224-4951-bf45-8d3dc99aa9d4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.145480] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] da8cf1ff-d977-4fd1-9239-043f07bed875/da8cf1ff-d977-4fd1-9239-043f07bed875.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.145747] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-475a955f-f05c-401c-8e4d-baa736c83505 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.167441] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 876.167441] env[63372]: value = "task-1024333" [ 876.167441] env[63372]: _type = "Task" [ 876.167441] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.175488] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024333, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.273517] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024331, 'name': ReconfigVM_Task, 'duration_secs': 0.612419} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.274086] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 1f52f9ee-2af3-4321-8ec6-5d5e841ed880/1f52f9ee-2af3-4321-8ec6-5d5e841ed880.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.274805] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-27e18624-5f70-477e-94c7-b12be89aeb59 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.283116] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 876.283116] env[63372]: value = "task-1024334" [ 876.283116] env[63372]: _type = "Task" [ 876.283116] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.293286] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024334, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.396970] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2e60ba7-9d8b-4d77-b1b3-3dad800ccf1b tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "ae6f3f42-7213-4ab1-b74c-1a557df6748b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.424s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.511815] env[63372]: DEBUG nova.network.neutron [-] [instance: ac90a156-be00-4f62-a76e-e08914531167] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.538273] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bf04129c-9e93-41ba-a0f6-1905215eefe0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.549759] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bfdfb15-e24d-48a2-ac1f-7389f83b9886 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.581508] env[63372]: DEBUG nova.compute.manager [req-4a7ddf2c-a694-4d75-80fe-127f39536504 req-be4e1316-3877-457b-9212-30443aa9be17 service nova] [instance: ac90a156-be00-4f62-a76e-e08914531167] Detach interface failed, port_id=40415ba1-4a9b-45d8-90ff-d95ac4e3980f, reason: Instance ac90a156-be00-4f62-a76e-e08914531167 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 876.631922] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f698aee-43d5-40e4-9f83-f38fba78fd7c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.640937] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd8530a-aa85-458d-bb67-b3e038bf5ca4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.678324] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d131e1d6-dbae-409f-9be4-5bf0e19d08a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.688567] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024333, 'name': ReconfigVM_Task, 'duration_secs': 0.259415} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.689825] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63284864-c9eb-48af-a0d8-f103d15f7e7d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.693605] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Reconfigured VM instance instance-0000004b to attach disk [datastore1] da8cf1ff-d977-4fd1-9239-043f07bed875/da8cf1ff-d977-4fd1-9239-043f07bed875.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.694210] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-260f176d-e1df-4e16-a6a4-053af5ff3908 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.706309] env[63372]: DEBUG nova.compute.provider_tree [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 876.708592] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 876.708592] env[63372]: value = "task-1024335" [ 876.708592] env[63372]: _type = "Task" [ 876.708592] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.717303] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024335, 'name': Rename_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.793695] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024334, 'name': Rename_Task, 'duration_secs': 0.205095} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.794040] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 876.794343] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96aae301-4ed7-4f1d-affd-cec07bf0bf33 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.801387] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 876.801387] env[63372]: value = "task-1024336" [ 876.801387] env[63372]: _type = "Task" [ 876.801387] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.811074] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024336, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.826812] env[63372]: DEBUG nova.compute.manager [req-ab626cba-3568-4aca-8ea4-14fe1ea3253a req-95e665b6-7fe9-4b6a-8ec9-7d0fea955908 service nova] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Received event network-vif-plugged-617ca099-5dd5-46c3-afc2-f80f4dbe242f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.827085] env[63372]: DEBUG oslo_concurrency.lockutils [req-ab626cba-3568-4aca-8ea4-14fe1ea3253a req-95e665b6-7fe9-4b6a-8ec9-7d0fea955908 service nova] Acquiring lock "c0b9bd7e-92e2-44a1-b461-e54c4411ddc8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.827330] env[63372]: DEBUG oslo_concurrency.lockutils [req-ab626cba-3568-4aca-8ea4-14fe1ea3253a req-95e665b6-7fe9-4b6a-8ec9-7d0fea955908 service nova] Lock "c0b9bd7e-92e2-44a1-b461-e54c4411ddc8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.827500] env[63372]: DEBUG oslo_concurrency.lockutils [req-ab626cba-3568-4aca-8ea4-14fe1ea3253a req-95e665b6-7fe9-4b6a-8ec9-7d0fea955908 service nova] Lock "c0b9bd7e-92e2-44a1-b461-e54c4411ddc8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.828233] env[63372]: DEBUG nova.compute.manager [req-ab626cba-3568-4aca-8ea4-14fe1ea3253a req-95e665b6-7fe9-4b6a-8ec9-7d0fea955908 service nova] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] No waiting events found dispatching network-vif-plugged-617ca099-5dd5-46c3-afc2-f80f4dbe242f {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 876.828233] env[63372]: WARNING nova.compute.manager [req-ab626cba-3568-4aca-8ea4-14fe1ea3253a req-95e665b6-7fe9-4b6a-8ec9-7d0fea955908 service nova] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Received unexpected event network-vif-plugged-617ca099-5dd5-46c3-afc2-f80f4dbe242f for instance with vm_state building and task_state spawning. [ 876.942094] env[63372]: DEBUG nova.network.neutron [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Successfully updated port: 617ca099-5dd5-46c3-afc2-f80f4dbe242f {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.014935] env[63372]: INFO nova.compute.manager [-] [instance: ac90a156-be00-4f62-a76e-e08914531167] Took 1.49 seconds to deallocate network for instance. [ 877.224715] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024335, 'name': Rename_Task, 'duration_secs': 0.277972} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.225202] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 877.225202] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4471073e-0887-4695-9ac0-24f5aa69e8db {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.233326] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 877.233326] env[63372]: value = "task-1024337" [ 877.233326] env[63372]: _type = "Task" [ 877.233326] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.245326] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024337, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.246823] env[63372]: DEBUG nova.scheduler.client.report [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 106 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 877.247132] env[63372]: DEBUG nova.compute.provider_tree [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 106 to 107 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 877.247334] env[63372]: DEBUG nova.compute.provider_tree [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 877.312944] env[63372]: DEBUG oslo_vmware.api [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024336, 'name': PowerOnVM_Task, 'duration_secs': 0.489383} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.313489] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 877.313714] env[63372]: INFO nova.compute.manager [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Took 8.60 seconds to spawn the instance on the hypervisor. [ 877.313900] env[63372]: DEBUG nova.compute.manager [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.314675] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75882ee1-9608-4986-b786-788edbc513f6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.446229] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "refresh_cache-c0b9bd7e-92e2-44a1-b461-e54c4411ddc8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.446376] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired lock "refresh_cache-c0b9bd7e-92e2-44a1-b461-e54c4411ddc8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.446583] env[63372]: DEBUG nova.network.neutron [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 877.523026] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.743754] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024337, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.752645] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.898s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.754793] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.240s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.755039] env[63372]: DEBUG nova.objects.instance [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lazy-loading 'resources' on Instance uuid cd0c01ac-602b-44a3-8099-84b8a50b2449 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.777845] env[63372]: INFO nova.scheduler.client.report [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Deleted allocations for instance 07dab964-a741-4723-98a3-0cfbe4eff975 [ 877.830722] env[63372]: INFO nova.compute.manager [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Took 34.66 seconds to build instance. [ 877.991041] env[63372]: DEBUG nova.network.neutron [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 878.247950] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024337, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.274507] env[63372]: INFO nova.compute.manager [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Rescuing [ 878.275080] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquiring lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.275080] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquired lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.275197] env[63372]: DEBUG nova.network.neutron [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 878.284969] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7cba4b6c-1107-4316-9967-316165d6fff3 tempest-ServerPasswordTestJSON-1877264766 tempest-ServerPasswordTestJSON-1877264766-project-member] Lock "07dab964-a741-4723-98a3-0cfbe4eff975" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.847s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.313288] env[63372]: DEBUG nova.network.neutron [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Updating instance_info_cache with network_info: [{"id": "617ca099-5dd5-46c3-afc2-f80f4dbe242f", "address": "fa:16:3e:e4:61:53", "network": {"id": "7157ae73-79e9-454e-8673-7841f4cfe1f5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-786035586-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4548b52cd704cc0b054e2f7d1562b9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap617ca099-5d", "ovs_interfaceid": "617ca099-5dd5-46c3-afc2-f80f4dbe242f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.332576] env[63372]: DEBUG oslo_concurrency.lockutils [None req-661fe728-8ac3-4d5f-84ae-3ae3a806ea20 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Lock "1f52f9ee-2af3-4321-8ec6-5d5e841ed880" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.174s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.528288] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d64d245a-9884-405a-bd06-bd4c9de45b9f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.536696] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb4e669e-c5f5-4eff-b9ed-906b82b77fd9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.569577] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c968ee-6f3a-4870-b0f2-fbf82da37079 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.577814] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-285f01e1-8fe8-4e95-9ca0-505922caafb4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.591983] env[63372]: DEBUG nova.compute.provider_tree [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.745529] env[63372]: DEBUG oslo_vmware.api [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024337, 'name': PowerOnVM_Task, 'duration_secs': 1.123252} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.745913] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 878.745913] env[63372]: DEBUG nova.compute.manager [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.746583] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9670d9-24ca-4bb6-ba24-168e93b77796 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.819079] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Releasing lock "refresh_cache-c0b9bd7e-92e2-44a1-b461-e54c4411ddc8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.819409] env[63372]: DEBUG nova.compute.manager [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Instance network_info: |[{"id": "617ca099-5dd5-46c3-afc2-f80f4dbe242f", "address": "fa:16:3e:e4:61:53", "network": {"id": "7157ae73-79e9-454e-8673-7841f4cfe1f5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-786035586-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4548b52cd704cc0b054e2f7d1562b9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap617ca099-5d", "ovs_interfaceid": "617ca099-5dd5-46c3-afc2-f80f4dbe242f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.819858] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:61:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '617ca099-5dd5-46c3-afc2-f80f4dbe242f', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.827303] env[63372]: DEBUG oslo.service.loopingcall [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.829906] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 878.830155] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-836edcce-8f4f-4a92-83ea-f84ae9a036a4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.850434] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.850434] env[63372]: value = "task-1024338" [ 878.850434] env[63372]: _type = "Task" [ 878.850434] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.852558] env[63372]: DEBUG nova.compute.manager [req-6665728b-84eb-4811-84e6-bb0c459efd1d req-8634ba1a-5b6e-4fd4-a08b-d58f0d4e4e63 service nova] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Received event network-changed-617ca099-5dd5-46c3-afc2-f80f4dbe242f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.852752] env[63372]: DEBUG nova.compute.manager [req-6665728b-84eb-4811-84e6-bb0c459efd1d req-8634ba1a-5b6e-4fd4-a08b-d58f0d4e4e63 service nova] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Refreshing instance network info cache due to event network-changed-617ca099-5dd5-46c3-afc2-f80f4dbe242f. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.852967] env[63372]: DEBUG oslo_concurrency.lockutils [req-6665728b-84eb-4811-84e6-bb0c459efd1d req-8634ba1a-5b6e-4fd4-a08b-d58f0d4e4e63 service nova] Acquiring lock "refresh_cache-c0b9bd7e-92e2-44a1-b461-e54c4411ddc8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.853129] env[63372]: DEBUG oslo_concurrency.lockutils [req-6665728b-84eb-4811-84e6-bb0c459efd1d req-8634ba1a-5b6e-4fd4-a08b-d58f0d4e4e63 service nova] Acquired lock "refresh_cache-c0b9bd7e-92e2-44a1-b461-e54c4411ddc8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.853293] env[63372]: DEBUG nova.network.neutron [req-6665728b-84eb-4811-84e6-bb0c459efd1d req-8634ba1a-5b6e-4fd4-a08b-d58f0d4e4e63 service nova] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Refreshing network info cache for port 617ca099-5dd5-46c3-afc2-f80f4dbe242f {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 878.863791] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024338, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.095137] env[63372]: DEBUG nova.scheduler.client.report [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.265179] env[63372]: DEBUG nova.network.neutron [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Updating instance_info_cache with network_info: [{"id": "75cfde99-66f3-4e6e-8755-061b37055d28", "address": "fa:16:3e:b9:fd:d5", "network": {"id": "00a725f6-e5e4-463e-a8ff-3a0953fd3186", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2027272115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "624e1cee468d4a538c70adf2e058869c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9950022e-0005-4da9-b1c4-6c9fec5ea180", "external-id": "nsx-vlan-transportzone-448", "segmentation_id": 448, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75cfde99-66", "ovs_interfaceid": "75cfde99-66f3-4e6e-8755-061b37055d28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.268506] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.363788] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024338, 'name': CreateVM_Task, 'duration_secs': 0.324619} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.364323] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 879.365015] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.365203] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.365523] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 879.365785] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7ea1c66-4db5-4988-8257-94d9ca4215e3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.371286] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 879.371286] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5257a70a-b6b0-3610-acff-0987aeacdac3" [ 879.371286] env[63372]: _type = "Task" [ 879.371286] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.382337] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5257a70a-b6b0-3610-acff-0987aeacdac3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.600729] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.846s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.603778] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.206s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.603778] env[63372]: DEBUG nova.objects.instance [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lazy-loading 'resources' on Instance uuid 798047e4-1a2b-458c-86c2-5567197d1861 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.629421] env[63372]: INFO nova.scheduler.client.report [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Deleted allocations for instance cd0c01ac-602b-44a3-8099-84b8a50b2449 [ 879.769749] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Releasing lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.882564] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5257a70a-b6b0-3610-acff-0987aeacdac3, 'name': SearchDatastore_Task, 'duration_secs': 0.013719} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.882742] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.882907] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.883157] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.883307] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.883487] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.883829] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7c7959c8-06ee-4189-993f-78c8bd0db956 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.894309] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.894609] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 879.895352] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d754f75-99be-48b5-9fb6-fd7cca544636 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.902097] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 879.902097] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5205402e-0543-1757-af61-10fae714a45d" [ 879.902097] env[63372]: _type = "Task" [ 879.902097] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.924914] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5205402e-0543-1757-af61-10fae714a45d, 'name': SearchDatastore_Task} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.924914] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8223174-fa60-44b6-a357-e08259612a1d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.934027] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 879.934027] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520027f2-318f-8657-7a90-6d16ed9f21b8" [ 879.934027] env[63372]: _type = "Task" [ 879.934027] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.944665] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520027f2-318f-8657-7a90-6d16ed9f21b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.130581] env[63372]: DEBUG nova.network.neutron [req-6665728b-84eb-4811-84e6-bb0c459efd1d req-8634ba1a-5b6e-4fd4-a08b-d58f0d4e4e63 service nova] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Updated VIF entry in instance network info cache for port 617ca099-5dd5-46c3-afc2-f80f4dbe242f. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 880.130937] env[63372]: DEBUG nova.network.neutron [req-6665728b-84eb-4811-84e6-bb0c459efd1d req-8634ba1a-5b6e-4fd4-a08b-d58f0d4e4e63 service nova] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Updating instance_info_cache with network_info: [{"id": "617ca099-5dd5-46c3-afc2-f80f4dbe242f", "address": "fa:16:3e:e4:61:53", "network": {"id": "7157ae73-79e9-454e-8673-7841f4cfe1f5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-786035586-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4548b52cd704cc0b054e2f7d1562b9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap617ca099-5d", "ovs_interfaceid": "617ca099-5dd5-46c3-afc2-f80f4dbe242f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.143469] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73f1037f-daac-4e6c-b244-d7eaf22af09a tempest-ServerRescueTestJSON-1856582621 tempest-ServerRescueTestJSON-1856582621-project-member] Lock "cd0c01ac-602b-44a3-8099-84b8a50b2449" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.855s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.303751] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 880.304098] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6af7b564-17a0-4d1e-9dc7-c302ae6b124f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.315651] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 880.315651] env[63372]: value = "task-1024339" [ 880.315651] env[63372]: _type = "Task" [ 880.315651] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.327078] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024339, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.362557] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb5cb75-85a1-4d5b-8eab-fd56fde89a62 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.370501] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4073f148-1eb6-4808-9eba-5aba586b8e28 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.405389] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a24a48-d593-48b1-a9d0-2b451a56c6bb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.414212] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a730f884-f6a0-4c25-87d6-ad6584057c97 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.428964] env[63372]: DEBUG nova.compute.provider_tree [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.440899] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520027f2-318f-8657-7a90-6d16ed9f21b8, 'name': SearchDatastore_Task, 'duration_secs': 0.011813} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.441788] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.442061] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] c0b9bd7e-92e2-44a1-b461-e54c4411ddc8/c0b9bd7e-92e2-44a1-b461-e54c4411ddc8.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 880.442310] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f77aad03-8989-4468-8cbe-2d17d8cb64a2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.454204] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 880.454204] env[63372]: value = "task-1024340" [ 880.454204] env[63372]: _type = "Task" [ 880.454204] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.463132] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024340, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.510432] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquiring lock "da8cf1ff-d977-4fd1-9239-043f07bed875" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.510733] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Lock "da8cf1ff-d977-4fd1-9239-043f07bed875" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.510969] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquiring lock "da8cf1ff-d977-4fd1-9239-043f07bed875-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.511245] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Lock "da8cf1ff-d977-4fd1-9239-043f07bed875-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.511463] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Lock "da8cf1ff-d977-4fd1-9239-043f07bed875-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.514068] env[63372]: INFO nova.compute.manager [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Terminating instance [ 880.516734] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquiring lock "refresh_cache-da8cf1ff-d977-4fd1-9239-043f07bed875" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.517014] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquired lock "refresh_cache-da8cf1ff-d977-4fd1-9239-043f07bed875" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.517300] env[63372]: DEBUG nova.network.neutron [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 880.636062] env[63372]: DEBUG oslo_concurrency.lockutils [req-6665728b-84eb-4811-84e6-bb0c459efd1d req-8634ba1a-5b6e-4fd4-a08b-d58f0d4e4e63 service nova] Releasing lock "refresh_cache-c0b9bd7e-92e2-44a1-b461-e54c4411ddc8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.829406] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024339, 'name': PowerOffVM_Task, 'duration_secs': 0.302155} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.829715] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 880.830541] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc52ec0-99ee-4b71-9a89-4998211f28e8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.852488] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb3b056-e63a-4ba6-ad0e-eb6e589e24b7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.896556] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 880.897134] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba7622fe-3455-413e-9907-dc3ee3d8ab9c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.907170] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 880.907170] env[63372]: value = "task-1024341" [ 880.907170] env[63372]: _type = "Task" [ 880.907170] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.916477] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] VM already powered off {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 880.916719] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 880.916976] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.917147] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.917340] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 880.917602] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a199e26f-6e09-4b28-9241-e3279c22bcc1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.925300] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 880.925488] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 880.926227] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4cd9bad-62f2-4032-b73e-0e5ce224c53a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.931522] env[63372]: DEBUG nova.scheduler.client.report [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 880.935453] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 880.935453] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5279e24d-814a-07af-3de3-0516cced2122" [ 880.935453] env[63372]: _type = "Task" [ 880.935453] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.943483] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5279e24d-814a-07af-3de3-0516cced2122, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.964641] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024340, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484657} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.964850] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] c0b9bd7e-92e2-44a1-b461-e54c4411ddc8/c0b9bd7e-92e2-44a1-b461-e54c4411ddc8.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 880.965412] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 880.965412] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d769c7e0-b044-4849-80d1-1f36477c085b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.971744] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 880.971744] env[63372]: value = "task-1024342" [ 880.971744] env[63372]: _type = "Task" [ 880.971744] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.979589] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024342, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.040214] env[63372]: DEBUG nova.network.neutron [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 881.090060] env[63372]: DEBUG nova.network.neutron [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.437400] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.834s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.440560] env[63372]: DEBUG oslo_concurrency.lockutils [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.921s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.440823] env[63372]: DEBUG nova.objects.instance [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lazy-loading 'resources' on Instance uuid 1d047728-50d8-465c-b217-6cbe9bbea3e8 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.452819] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5279e24d-814a-07af-3de3-0516cced2122, 'name': SearchDatastore_Task, 'duration_secs': 0.011446} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.453816] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf6cadf8-b868-452b-a326-8edf68963f5a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.459446] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 881.459446] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52ae3b6f-1f61-1980-29f2-92f5327f2ff5" [ 881.459446] env[63372]: _type = "Task" [ 881.459446] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.463531] env[63372]: INFO nova.scheduler.client.report [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Deleted allocations for instance 798047e4-1a2b-458c-86c2-5567197d1861 [ 881.470073] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ae3b6f-1f61-1980-29f2-92f5327f2ff5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.481995] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024342, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063135} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.482274] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.483524] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c0a7a6-b9fb-4070-86a8-dae9041146cc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.505106] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] c0b9bd7e-92e2-44a1-b461-e54c4411ddc8/c0b9bd7e-92e2-44a1-b461-e54c4411ddc8.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.506042] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1af41cc3-3fba-4d08-9377-caa09aeef41b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.526812] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 881.526812] env[63372]: value = "task-1024343" [ 881.526812] env[63372]: _type = "Task" [ 881.526812] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.535050] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024343, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.592304] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Releasing lock "refresh_cache-da8cf1ff-d977-4fd1-9239-043f07bed875" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.592646] env[63372]: DEBUG nova.compute.manager [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 881.592853] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 881.593728] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da714e2d-4fcd-49e2-9557-29264d9fe615 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.601578] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 881.601862] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4892deba-e38a-4817-8774-effc7f9a6ce0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.611564] env[63372]: DEBUG oslo_vmware.api [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 881.611564] env[63372]: value = "task-1024344" [ 881.611564] env[63372]: _type = "Task" [ 881.611564] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.623144] env[63372]: DEBUG oslo_vmware.api [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024344, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.977504] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ae3b6f-1f61-1980-29f2-92f5327f2ff5, 'name': SearchDatastore_Task, 'duration_secs': 0.016867} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.978343] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2c1c3b9-648a-4b30-8723-436e66134b99 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "798047e4-1a2b-458c-86c2-5567197d1861" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.898s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.979474] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.979857] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 1f52f9ee-2af3-4321-8ec6-5d5e841ed880/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk. {{(pid=63372) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 881.980448] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e79c04cd-4d15-4eb7-a756-673905dd6517 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.993166] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 881.993166] env[63372]: value = "task-1024345" [ 881.993166] env[63372]: _type = "Task" [ 881.993166] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.000646] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024345, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.039639] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024343, 'name': ReconfigVM_Task, 'duration_secs': 0.287023} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.040867] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Reconfigured VM instance instance-0000004d to attach disk [datastore2] c0b9bd7e-92e2-44a1-b461-e54c4411ddc8/c0b9bd7e-92e2-44a1-b461-e54c4411ddc8.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.040867] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2d56289d-1546-4c0d-a8e0-2070827ec2b4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.051284] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 882.051284] env[63372]: value = "task-1024346" [ 882.051284] env[63372]: _type = "Task" [ 882.051284] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.059989] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024346, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.123497] env[63372]: DEBUG oslo_vmware.api [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024344, 'name': PowerOffVM_Task, 'duration_secs': 0.117203} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.123745] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 882.123916] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 882.124185] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-27899521-3c88-4eed-b9bb-9362340851c0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.150395] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 882.150655] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 882.150911] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Deleting the datastore file [datastore1] da8cf1ff-d977-4fd1-9239-043f07bed875 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.153659] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e6a092e-c978-452f-9b22-d3f0a9c4510b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.160399] env[63372]: DEBUG oslo_vmware.api [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for the task: (returnval){ [ 882.160399] env[63372]: value = "task-1024348" [ 882.160399] env[63372]: _type = "Task" [ 882.160399] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.168422] env[63372]: DEBUG oslo_vmware.api [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024348, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.203160] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb1a4ed-43e6-42d1-90da-2ffe3f27192d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.210206] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3bdada5-d913-4553-95b7-12e53790499f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.249764] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d7c8cb-9084-41eb-b928-1e6ec9257d8d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.258524] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5c98ca-2aa7-44c9-99d3-61e1eb698e17 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.275592] env[63372]: DEBUG nova.compute.provider_tree [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.501253] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024345, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488178} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.501520] env[63372]: INFO nova.virt.vmwareapi.ds_util [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 1f52f9ee-2af3-4321-8ec6-5d5e841ed880/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk. [ 882.502350] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313d64b5-666f-4401-ae96-46e893b0131a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.527078] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 1f52f9ee-2af3-4321-8ec6-5d5e841ed880/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.527354] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d9eb6c9-1b56-4918-8144-947e55c3eebc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.547253] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 882.547253] env[63372]: value = "task-1024349" [ 882.547253] env[63372]: _type = "Task" [ 882.547253] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.557474] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024349, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.562800] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024346, 'name': Rename_Task, 'duration_secs': 0.145763} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.563015] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 882.563280] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3f087f6-60b3-40cf-a596-870b2c0b9481 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.569286] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 882.569286] env[63372]: value = "task-1024350" [ 882.569286] env[63372]: _type = "Task" [ 882.569286] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.577030] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024350, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.669831] env[63372]: DEBUG oslo_vmware.api [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Task: {'id': task-1024348, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194821} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.670141] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.670342] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 882.670523] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 882.670697] env[63372]: INFO nova.compute.manager [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Took 1.08 seconds to destroy the instance on the hypervisor. [ 882.671065] env[63372]: DEBUG oslo.service.loopingcall [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.671258] env[63372]: DEBUG nova.compute.manager [-] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 882.671307] env[63372]: DEBUG nova.network.neutron [-] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 882.686349] env[63372]: DEBUG nova.network.neutron [-] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 882.779348] env[63372]: DEBUG nova.scheduler.client.report [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 883.058090] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024349, 'name': ReconfigVM_Task, 'duration_secs': 0.296126} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.058426] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 1f52f9ee-2af3-4321-8ec6-5d5e841ed880/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.059316] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7458ef1-6ab2-47ec-a73e-696403e77d14 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.085980] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1145604-3945-4003-97c3-a512939edb94 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.103070] env[63372]: DEBUG oslo_vmware.api [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024350, 'name': PowerOnVM_Task, 'duration_secs': 0.461322} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.104308] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 883.104518] env[63372]: INFO nova.compute.manager [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Took 7.13 seconds to spawn the instance on the hypervisor. [ 883.104911] env[63372]: DEBUG nova.compute.manager [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.105306] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 883.105306] env[63372]: value = "task-1024351" [ 883.105306] env[63372]: _type = "Task" [ 883.105306] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.106079] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efbf95f9-c298-4ea4-8e05-f3a9aaed296a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.120138] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024351, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.189549] env[63372]: DEBUG nova.network.neutron [-] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.284676] env[63372]: DEBUG oslo_concurrency.lockutils [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.844s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.287204] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 21.392s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.310323] env[63372]: INFO nova.scheduler.client.report [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Deleted allocations for instance 1d047728-50d8-465c-b217-6cbe9bbea3e8 [ 883.618628] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024351, 'name': ReconfigVM_Task, 'duration_secs': 0.19227} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.619017] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 883.619414] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b52d1d12-f0b7-484f-bb87-3c56ba13b796 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.626864] env[63372]: INFO nova.compute.manager [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Took 34.91 seconds to build instance. [ 883.629584] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 883.629584] env[63372]: value = "task-1024352" [ 883.629584] env[63372]: _type = "Task" [ 883.629584] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.638408] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024352, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.692644] env[63372]: INFO nova.compute.manager [-] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Took 1.02 seconds to deallocate network for instance. [ 883.792055] env[63372]: INFO nova.compute.claims [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 883.817050] env[63372]: DEBUG oslo_concurrency.lockutils [None req-97ca507b-2182-40f0-9117-730868ce4325 tempest-MultipleCreateTestJSON-1527533804 tempest-MultipleCreateTestJSON-1527533804-project-member] Lock "1d047728-50d8-465c-b217-6cbe9bbea3e8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.214s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.131370] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0a117a77-741d-4fd8-90dc-f708ebc1a9f5 tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "c0b9bd7e-92e2-44a1-b461-e54c4411ddc8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.429s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.142150] env[63372]: DEBUG oslo_vmware.api [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024352, 'name': PowerOnVM_Task, 'duration_secs': 0.402173} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.143367] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 884.146422] env[63372]: DEBUG nova.compute.manager [None req-9bcfe0a9-6226-4db3-93b5-18144cc59cfe tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.147286] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63926e41-958b-434f-834b-5ce10f6dae87 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.198427] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.300309] env[63372]: INFO nova.compute.resource_tracker [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating resource usage from migration 00ae52e3-01cb-4775-870a-c0565c40be33 [ 884.537053] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54566c1-4906-46fb-aa6d-b7d2f9bbbb25 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.546631] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb42d26a-a3ed-4bcb-ac1e-50ed155c3d29 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.581013] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fed5f20-a2d6-4d4a-8f64-d7966fc33832 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.588870] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f580a3-0f91-4715-8159-3927f618ba19 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.602696] env[63372]: DEBUG nova.compute.provider_tree [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 885.148121] env[63372]: DEBUG nova.scheduler.client.report [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 107 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 885.148428] env[63372]: DEBUG nova.compute.provider_tree [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 107 to 108 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 885.148827] env[63372]: DEBUG nova.compute.provider_tree [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 885.531094] env[63372]: DEBUG nova.compute.manager [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.531988] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aadc977-f0ee-4db0-a9c2-d7ae6d671164 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.654445] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.367s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.654670] env[63372]: INFO nova.compute.manager [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Migrating [ 885.661385] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.400s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.665021] env[63372]: DEBUG nova.objects.instance [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lazy-loading 'resources' on Instance uuid c01a5d24-eb46-4a69-993e-753880ce8e85 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.727818] env[63372]: DEBUG nova.compute.manager [req-f6bfc2b5-b3e4-4d60-9634-ba2f4362f136 req-58fbc6bd-eb83-43c5-a2e0-a7fe82dea39f service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Received event network-changed-75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.728213] env[63372]: DEBUG nova.compute.manager [req-f6bfc2b5-b3e4-4d60-9634-ba2f4362f136 req-58fbc6bd-eb83-43c5-a2e0-a7fe82dea39f service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Refreshing instance network info cache due to event network-changed-75cfde99-66f3-4e6e-8755-061b37055d28. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 885.728286] env[63372]: DEBUG oslo_concurrency.lockutils [req-f6bfc2b5-b3e4-4d60-9634-ba2f4362f136 req-58fbc6bd-eb83-43c5-a2e0-a7fe82dea39f service nova] Acquiring lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.728413] env[63372]: DEBUG oslo_concurrency.lockutils [req-f6bfc2b5-b3e4-4d60-9634-ba2f4362f136 req-58fbc6bd-eb83-43c5-a2e0-a7fe82dea39f service nova] Acquired lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.728578] env[63372]: DEBUG nova.network.neutron [req-f6bfc2b5-b3e4-4d60-9634-ba2f4362f136 req-58fbc6bd-eb83-43c5-a2e0-a7fe82dea39f service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Refreshing network info cache for port 75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 886.046157] env[63372]: INFO nova.compute.manager [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] instance snapshotting [ 886.060215] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5831cb8b-b9f4-41db-adda-d075a1dede8e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.073183] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0c5ab2-4248-426b-bb5c-5e189d8dae28 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.165780] env[63372]: DEBUG nova.objects.instance [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lazy-loading 'numa_topology' on Instance uuid c01a5d24-eb46-4a69-993e-753880ce8e85 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.171736] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.171736] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.171736] env[63372]: DEBUG nova.network.neutron [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 886.575164] env[63372]: DEBUG nova.network.neutron [req-f6bfc2b5-b3e4-4d60-9634-ba2f4362f136 req-58fbc6bd-eb83-43c5-a2e0-a7fe82dea39f service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Updated VIF entry in instance network info cache for port 75cfde99-66f3-4e6e-8755-061b37055d28. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 886.575551] env[63372]: DEBUG nova.network.neutron [req-f6bfc2b5-b3e4-4d60-9634-ba2f4362f136 req-58fbc6bd-eb83-43c5-a2e0-a7fe82dea39f service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Updating instance_info_cache with network_info: [{"id": "75cfde99-66f3-4e6e-8755-061b37055d28", "address": "fa:16:3e:b9:fd:d5", "network": {"id": "00a725f6-e5e4-463e-a8ff-3a0953fd3186", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2027272115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "624e1cee468d4a538c70adf2e058869c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9950022e-0005-4da9-b1c4-6c9fec5ea180", "external-id": "nsx-vlan-transportzone-448", "segmentation_id": 448, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75cfde99-66", "ovs_interfaceid": "75cfde99-66f3-4e6e-8755-061b37055d28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.585063] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 886.585488] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-fd289d9c-58c9-48e7-a722-75369bbeec40 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.594958] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 886.594958] env[63372]: value = "task-1024353" [ 886.594958] env[63372]: _type = "Task" [ 886.594958] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.604219] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024353, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.668354] env[63372]: DEBUG nova.objects.base [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 886.915325] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd65b3b-d3fa-4262-9873-19345fdad4c2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.922674] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66670ff5-a036-4eff-954b-82947f0caef1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.957992] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c69e9922-0375-456e-bfff-1e56efbbc8a4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.965788] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236bff7a-e5c1-4c95-88a3-847bbb79aa26 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.979294] env[63372]: DEBUG nova.compute.provider_tree [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.079200] env[63372]: DEBUG nova.network.neutron [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance_info_cache with network_info: [{"id": "e1486286-9147-4e5e-88c1-60893a3338dd", "address": "fa:16:3e:6c:92:ac", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1486286-91", "ovs_interfaceid": "e1486286-9147-4e5e-88c1-60893a3338dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.080529] env[63372]: DEBUG oslo_concurrency.lockutils [req-f6bfc2b5-b3e4-4d60-9634-ba2f4362f136 req-58fbc6bd-eb83-43c5-a2e0-a7fe82dea39f service nova] Releasing lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.104709] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024353, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.482338] env[63372]: DEBUG nova.scheduler.client.report [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 887.582512] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.605710] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024353, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.988032] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.326s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.991340] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.780s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.993358] env[63372]: INFO nova.compute.claims [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 888.109795] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024353, 'name': CreateSnapshot_Task, 'duration_secs': 1.343397} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.110169] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 888.110966] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ace5b55-99dd-42aa-a9f2-8f73e1ed4d74 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.282751] env[63372]: DEBUG nova.compute.manager [req-95954fd6-07e0-47c4-b6b6-dc6351a74f2f req-7bcfa5ec-fdae-4c62-b41c-fd7c57eb83fe service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Received event network-changed-75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.283086] env[63372]: DEBUG nova.compute.manager [req-95954fd6-07e0-47c4-b6b6-dc6351a74f2f req-7bcfa5ec-fdae-4c62-b41c-fd7c57eb83fe service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Refreshing instance network info cache due to event network-changed-75cfde99-66f3-4e6e-8755-061b37055d28. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 888.283262] env[63372]: DEBUG oslo_concurrency.lockutils [req-95954fd6-07e0-47c4-b6b6-dc6351a74f2f req-7bcfa5ec-fdae-4c62-b41c-fd7c57eb83fe service nova] Acquiring lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.283414] env[63372]: DEBUG oslo_concurrency.lockutils [req-95954fd6-07e0-47c4-b6b6-dc6351a74f2f req-7bcfa5ec-fdae-4c62-b41c-fd7c57eb83fe service nova] Acquired lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.283580] env[63372]: DEBUG nova.network.neutron [req-95954fd6-07e0-47c4-b6b6-dc6351a74f2f req-7bcfa5ec-fdae-4c62-b41c-fd7c57eb83fe service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Refreshing network info cache for port 75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 888.503240] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d4469a59-e4f7-4258-847c-bf327de3515c tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 46.418s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.504685] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 22.897s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.504896] env[63372]: INFO nova.compute.manager [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Unshelving [ 888.629535] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 888.629858] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8be5fced-3d59-40f4-a19d-30b0d43c931c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.640576] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 888.640576] env[63372]: value = "task-1024354" [ 888.640576] env[63372]: _type = "Task" [ 888.640576] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.648894] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024354, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.999237] env[63372]: DEBUG nova.network.neutron [req-95954fd6-07e0-47c4-b6b6-dc6351a74f2f req-7bcfa5ec-fdae-4c62-b41c-fd7c57eb83fe service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Updated VIF entry in instance network info cache for port 75cfde99-66f3-4e6e-8755-061b37055d28. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 888.999677] env[63372]: DEBUG nova.network.neutron [req-95954fd6-07e0-47c4-b6b6-dc6351a74f2f req-7bcfa5ec-fdae-4c62-b41c-fd7c57eb83fe service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Updating instance_info_cache with network_info: [{"id": "75cfde99-66f3-4e6e-8755-061b37055d28", "address": "fa:16:3e:b9:fd:d5", "network": {"id": "00a725f6-e5e4-463e-a8ff-3a0953fd3186", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2027272115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "624e1cee468d4a538c70adf2e058869c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9950022e-0005-4da9-b1c4-6c9fec5ea180", "external-id": "nsx-vlan-transportzone-448", "segmentation_id": 448, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75cfde99-66", "ovs_interfaceid": "75cfde99-66f3-4e6e-8755-061b37055d28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.096741] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c7a44b-8ef0-47a8-b7d9-f96000a66bb4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.117704] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance '19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5' progress to 0 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 889.152550] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024354, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.248083] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2122b1fe-4145-4ad5-badd-45c45690e5a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.257868] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ca2dea-7235-4b2c-a4d6-9ad6721f909f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.289160] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c93f039-9df4-48d3-a913-b5b91ff4daa8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.296410] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a308529d-cd61-4b72-ab17-a5bf21e3dfad {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.309761] env[63372]: DEBUG nova.compute.provider_tree [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.502240] env[63372]: DEBUG oslo_concurrency.lockutils [req-95954fd6-07e0-47c4-b6b6-dc6351a74f2f req-7bcfa5ec-fdae-4c62-b41c-fd7c57eb83fe service nova] Releasing lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.515232] env[63372]: DEBUG nova.compute.utils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 889.626292] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 889.626665] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b062078a-75ce-41e9-85c4-abe6d41aabc0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.633199] env[63372]: DEBUG oslo_vmware.api [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 889.633199] env[63372]: value = "task-1024355" [ 889.633199] env[63372]: _type = "Task" [ 889.633199] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.641460] env[63372]: DEBUG oslo_vmware.api [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024355, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.649711] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024354, 'name': CloneVM_Task, 'duration_secs': 0.972684} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.650176] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Created linked-clone VM from snapshot [ 889.650645] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9fdaeba-0a7a-4bb4-a080-ac1dea93c18e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.657356] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Uploading image 98f6633f-5f4e-473c-8fc9-84d6976c7c02 {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 889.691576] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 889.691576] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-88a98664-ba19-471d-b3b5-79b515519a22 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.696210] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 889.696210] env[63372]: value = "task-1024356" [ 889.696210] env[63372]: _type = "Task" [ 889.696210] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.704548] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024356, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.815175] env[63372]: DEBUG nova.scheduler.client.report [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.018441] env[63372]: INFO nova.virt.block_device [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Booting with volume cccbd23a-5f23-47d3-9e07-0c6321e2cfe3 at /dev/sdb [ 890.082816] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9841c125-faba-48e3-87e1-c59f28dbd566 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.091830] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-930477bc-f2cd-4056-a08c-1fcf8043cc8a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.122333] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a3850de3-fe35-419d-a153-e0dcdf6d4243 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.131034] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50326cd-379e-434d-8d69-bc5f3a5c0ed9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.150500] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] VM already powered off {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 890.150722] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance '19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5' progress to 17 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 890.161291] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2046dd9-6ce1-423c-b7d3-63b1fb84fc2b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.170072] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76310786-2b52-471f-b74c-e87af2678953 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.181942] env[63372]: DEBUG nova.virt.block_device [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updating existing volume attachment record: b2bc4986-aba4-402c-8024-89c05a6c57e7 {{(pid=63372) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 890.208238] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024356, 'name': Destroy_Task, 'duration_secs': 0.416197} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.208626] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Destroyed the VM [ 890.209122] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 890.209355] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-29c527ce-7e9d-41df-92df-19e0d33a0e6f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.217569] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 890.217569] env[63372]: value = "task-1024357" [ 890.217569] env[63372]: _type = "Task" [ 890.217569] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.229874] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024357, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.319138] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.328s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.319819] env[63372]: DEBUG nova.compute.manager [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 890.324177] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.306s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.329020] env[63372]: INFO nova.compute.claims [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 890.487196] env[63372]: DEBUG nova.compute.manager [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Received event network-changed-75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.487619] env[63372]: DEBUG nova.compute.manager [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Refreshing instance network info cache due to event network-changed-75cfde99-66f3-4e6e-8755-061b37055d28. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 890.487912] env[63372]: DEBUG oslo_concurrency.lockutils [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] Acquiring lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.488113] env[63372]: DEBUG oslo_concurrency.lockutils [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] Acquired lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.488401] env[63372]: DEBUG nova.network.neutron [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Refreshing network info cache for port 75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 890.663674] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 890.664035] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 890.664119] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 890.664291] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 890.664442] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 890.664594] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 890.664799] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 890.664961] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 890.665161] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 890.665328] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 890.665505] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 890.670727] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e07fe0c6-8d9a-4b34-9746-14aeef974691 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.686676] env[63372]: DEBUG oslo_vmware.api [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 890.686676] env[63372]: value = "task-1024361" [ 890.686676] env[63372]: _type = "Task" [ 890.686676] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.694759] env[63372]: DEBUG oslo_vmware.api [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024361, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.726917] env[63372]: DEBUG oslo_vmware.api [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024357, 'name': RemoveSnapshot_Task, 'duration_secs': 0.501735} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.727211] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 890.832303] env[63372]: DEBUG nova.compute.utils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 890.836275] env[63372]: DEBUG nova.compute.manager [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 890.836472] env[63372]: DEBUG nova.network.neutron [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 890.963589] env[63372]: DEBUG nova.policy [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '881df5e623b340a487f37076c0cd4a15', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b6afb5579f2b48e4b094f6490136417f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 891.186491] env[63372]: DEBUG nova.network.neutron [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Updated VIF entry in instance network info cache for port 75cfde99-66f3-4e6e-8755-061b37055d28. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 891.186881] env[63372]: DEBUG nova.network.neutron [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Updating instance_info_cache with network_info: [{"id": "75cfde99-66f3-4e6e-8755-061b37055d28", "address": "fa:16:3e:b9:fd:d5", "network": {"id": "00a725f6-e5e4-463e-a8ff-3a0953fd3186", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2027272115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "624e1cee468d4a538c70adf2e058869c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9950022e-0005-4da9-b1c4-6c9fec5ea180", "external-id": "nsx-vlan-transportzone-448", "segmentation_id": 448, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75cfde99-66", "ovs_interfaceid": "75cfde99-66f3-4e6e-8755-061b37055d28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.197607] env[63372]: DEBUG oslo_vmware.api [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024361, 'name': ReconfigVM_Task, 'duration_secs': 0.168875} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.198670] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance '19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5' progress to 33 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 891.232243] env[63372]: WARNING nova.compute.manager [None req-757bbd24-4d3d-4b73-8c52-fbf7226c70df tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Image not found during snapshot: nova.exception.ImageNotFound: Image 98f6633f-5f4e-473c-8fc9-84d6976c7c02 could not be found. [ 891.331907] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquiring lock "1f52f9ee-2af3-4321-8ec6-5d5e841ed880" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.332260] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Lock "1f52f9ee-2af3-4321-8ec6-5d5e841ed880" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.332487] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquiring lock "1f52f9ee-2af3-4321-8ec6-5d5e841ed880-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.332677] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Lock "1f52f9ee-2af3-4321-8ec6-5d5e841ed880-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.332860] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Lock "1f52f9ee-2af3-4321-8ec6-5d5e841ed880-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.335609] env[63372]: DEBUG nova.compute.manager [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 891.339433] env[63372]: INFO nova.compute.manager [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Terminating instance [ 891.345192] env[63372]: DEBUG nova.compute.manager [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 891.345402] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 891.346393] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c395254d-4a90-48f5-ba8f-5ea5c1f120b1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.355510] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 891.355742] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f75fea69-257d-4c7c-bffc-20a29012893e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.363837] env[63372]: DEBUG oslo_vmware.api [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 891.363837] env[63372]: value = "task-1024362" [ 891.363837] env[63372]: _type = "Task" [ 891.363837] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.372224] env[63372]: DEBUG oslo_vmware.api [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024362, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.429180] env[63372]: DEBUG nova.network.neutron [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Successfully created port: 777f211d-71d8-4f32-bfac-588b17d506f9 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 891.601403] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39563e61-d619-4e00-bf43-d5f6449f5741 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.615186] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a259b65-e240-4ecb-85db-ce105f654836 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.650055] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-216e5263-1250-4ace-8e9b-a96f5f85494c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.659924] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e65200-b7d0-4a50-8f76-563f142d7f15 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.674830] env[63372]: DEBUG nova.compute.provider_tree [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.693504] env[63372]: DEBUG oslo_concurrency.lockutils [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] Releasing lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.693504] env[63372]: DEBUG nova.compute.manager [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Received event network-changed-75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.693643] env[63372]: DEBUG nova.compute.manager [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Refreshing instance network info cache due to event network-changed-75cfde99-66f3-4e6e-8755-061b37055d28. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 891.693840] env[63372]: DEBUG oslo_concurrency.lockutils [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] Acquiring lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.693990] env[63372]: DEBUG oslo_concurrency.lockutils [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] Acquired lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.694174] env[63372]: DEBUG nova.network.neutron [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Refreshing network info cache for port 75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 891.708332] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 891.708591] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 891.708753] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 891.709027] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 891.709183] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 891.709553] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 891.709619] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 891.709768] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 891.710051] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 891.710253] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 891.710435] env[63372]: DEBUG nova.virt.hardware [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 891.715697] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Reconfiguring VM instance instance-00000024 to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 891.716237] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9ca4b75-a845-4ba6-90c3-4d1de1987b63 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.733945] env[63372]: DEBUG oslo_vmware.api [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 891.733945] env[63372]: value = "task-1024363" [ 891.733945] env[63372]: _type = "Task" [ 891.733945] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.742100] env[63372]: DEBUG oslo_vmware.api [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024363, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.801489] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "c0b9bd7e-92e2-44a1-b461-e54c4411ddc8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.801761] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "c0b9bd7e-92e2-44a1-b461-e54c4411ddc8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.801976] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "c0b9bd7e-92e2-44a1-b461-e54c4411ddc8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.802180] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "c0b9bd7e-92e2-44a1-b461-e54c4411ddc8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.802358] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "c0b9bd7e-92e2-44a1-b461-e54c4411ddc8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.804729] env[63372]: INFO nova.compute.manager [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Terminating instance [ 891.806516] env[63372]: DEBUG nova.compute.manager [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 891.806713] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 891.807622] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-501826f0-11a6-405f-937d-97793709e448 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.815111] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 891.815348] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c9d2cb1a-b275-4770-8af9-b1f5a204fa56 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.822225] env[63372]: DEBUG oslo_vmware.api [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 891.822225] env[63372]: value = "task-1024364" [ 891.822225] env[63372]: _type = "Task" [ 891.822225] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.829417] env[63372]: DEBUG oslo_vmware.api [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024364, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.872104] env[63372]: DEBUG oslo_vmware.api [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024362, 'name': PowerOffVM_Task, 'duration_secs': 0.277528} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.872439] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 891.872622] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 891.872923] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4aa081fe-b55d-4cc3-813a-70fe62a53c61 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.943772] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 891.943772] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 891.943953] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Deleting the datastore file [datastore2] 1f52f9ee-2af3-4321-8ec6-5d5e841ed880 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.944496] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b37f2801-c214-4b68-94b0-9a7c27002ef8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.951594] env[63372]: DEBUG oslo_vmware.api [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for the task: (returnval){ [ 891.951594] env[63372]: value = "task-1024366" [ 891.951594] env[63372]: _type = "Task" [ 891.951594] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.961326] env[63372]: DEBUG oslo_vmware.api [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024366, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.178529] env[63372]: DEBUG nova.scheduler.client.report [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 892.246360] env[63372]: DEBUG oslo_vmware.api [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024363, 'name': ReconfigVM_Task, 'duration_secs': 0.155029} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.246640] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Reconfigured VM instance instance-00000024 to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 892.247480] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658ef852-855f-45bb-ab6e-c366f3ac2baf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.270364] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5/19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.271070] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88d2a341-dbbe-469e-9892-0d2c803aafbb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.291264] env[63372]: DEBUG oslo_vmware.api [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 892.291264] env[63372]: value = "task-1024367" [ 892.291264] env[63372]: _type = "Task" [ 892.291264] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.299423] env[63372]: DEBUG oslo_vmware.api [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024367, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.332340] env[63372]: DEBUG oslo_vmware.api [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024364, 'name': PowerOffVM_Task, 'duration_secs': 0.225859} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.332340] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 892.332340] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 892.332597] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0e558f3a-5601-4a1e-9a2c-e1227704cac4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.343485] env[63372]: DEBUG nova.compute.manager [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 892.395212] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 892.395596] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 892.395897] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Deleting the datastore file [datastore2] c0b9bd7e-92e2-44a1-b461-e54c4411ddc8 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 892.398356] env[63372]: DEBUG nova.virt.hardware [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 892.398625] env[63372]: DEBUG nova.virt.hardware [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 892.398881] env[63372]: DEBUG nova.virt.hardware [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 892.399116] env[63372]: DEBUG nova.virt.hardware [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 892.399304] env[63372]: DEBUG nova.virt.hardware [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 892.399492] env[63372]: DEBUG nova.virt.hardware [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 892.399767] env[63372]: DEBUG nova.virt.hardware [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 892.399962] env[63372]: DEBUG nova.virt.hardware [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 892.400188] env[63372]: DEBUG nova.virt.hardware [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 892.400402] env[63372]: DEBUG nova.virt.hardware [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 892.400621] env[63372]: DEBUG nova.virt.hardware [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 892.400921] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ce9840cb-ce64-4ae7-8fa0-85b29a4fe71b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.403281] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e804557-24ec-4d2c-ab1d-a774c9d26def {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.411959] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7048b0b-ad1d-41d2-bf55-f3095c1edda8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.415909] env[63372]: DEBUG oslo_vmware.api [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for the task: (returnval){ [ 892.415909] env[63372]: value = "task-1024369" [ 892.415909] env[63372]: _type = "Task" [ 892.415909] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.416695] env[63372]: DEBUG nova.network.neutron [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Updated VIF entry in instance network info cache for port 75cfde99-66f3-4e6e-8755-061b37055d28. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 892.417085] env[63372]: DEBUG nova.network.neutron [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Updating instance_info_cache with network_info: [{"id": "75cfde99-66f3-4e6e-8755-061b37055d28", "address": "fa:16:3e:b9:fd:d5", "network": {"id": "00a725f6-e5e4-463e-a8ff-3a0953fd3186", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2027272115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "624e1cee468d4a538c70adf2e058869c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9950022e-0005-4da9-b1c4-6c9fec5ea180", "external-id": "nsx-vlan-transportzone-448", "segmentation_id": 448, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75cfde99-66", "ovs_interfaceid": "75cfde99-66f3-4e6e-8755-061b37055d28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.435102] env[63372]: DEBUG oslo_vmware.api [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024369, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.459523] env[63372]: DEBUG oslo_vmware.api [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Task: {'id': task-1024366, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174026} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.459746] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.459856] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 892.460049] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 892.460229] env[63372]: INFO nova.compute.manager [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Took 1.11 seconds to destroy the instance on the hypervisor. [ 892.460470] env[63372]: DEBUG oslo.service.loopingcall [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.460659] env[63372]: DEBUG nova.compute.manager [-] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 892.460784] env[63372]: DEBUG nova.network.neutron [-] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 892.684793] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.360s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.685173] env[63372]: DEBUG nova.compute.manager [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 892.687420] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.463s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.688870] env[63372]: INFO nova.compute.claims [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 892.800862] env[63372]: DEBUG oslo_vmware.api [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024367, 'name': ReconfigVM_Task, 'duration_secs': 0.28115} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.801173] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Reconfigured VM instance instance-00000024 to attach disk [datastore2] 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5/19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.801495] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance '19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5' progress to 50 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 892.887615] env[63372]: DEBUG nova.compute.manager [req-fa1f3085-eaf1-4bee-8b9a-83e8300f5170 req-5dd8373a-f2f3-41d7-a201-391f4714845e service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Received event network-vif-deleted-75cfde99-66f3-4e6e-8755-061b37055d28 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.888312] env[63372]: INFO nova.compute.manager [req-fa1f3085-eaf1-4bee-8b9a-83e8300f5170 req-5dd8373a-f2f3-41d7-a201-391f4714845e service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Neutron deleted interface 75cfde99-66f3-4e6e-8755-061b37055d28; detaching it from the instance and deleting it from the info cache [ 892.888516] env[63372]: DEBUG nova.network.neutron [req-fa1f3085-eaf1-4bee-8b9a-83e8300f5170 req-5dd8373a-f2f3-41d7-a201-391f4714845e service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.924314] env[63372]: DEBUG oslo_concurrency.lockutils [req-b88ec4a4-5be9-4973-81a9-cdc6a76cdd3b req-a16f66cd-f709-47b1-ac56-75103327fef3 service nova] Releasing lock "refresh_cache-1f52f9ee-2af3-4321-8ec6-5d5e841ed880" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.930117] env[63372]: DEBUG oslo_vmware.api [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Task: {'id': task-1024369, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128975} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.930349] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.930535] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 892.930710] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 892.930879] env[63372]: INFO nova.compute.manager [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Took 1.12 seconds to destroy the instance on the hypervisor. [ 892.931141] env[63372]: DEBUG oslo.service.loopingcall [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.931338] env[63372]: DEBUG nova.compute.manager [-] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 892.931430] env[63372]: DEBUG nova.network.neutron [-] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 893.121248] env[63372]: DEBUG nova.compute.manager [req-e1037469-0611-4760-bcbf-caa3a66c8b52 req-d435ec17-d780-4ac8-966a-e1ba29e9d2ea service nova] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Received event network-vif-plugged-777f211d-71d8-4f32-bfac-588b17d506f9 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 893.121504] env[63372]: DEBUG oslo_concurrency.lockutils [req-e1037469-0611-4760-bcbf-caa3a66c8b52 req-d435ec17-d780-4ac8-966a-e1ba29e9d2ea service nova] Acquiring lock "382fa76b-b930-459e-b84e-da1b9eb74b74-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.122186] env[63372]: DEBUG oslo_concurrency.lockutils [req-e1037469-0611-4760-bcbf-caa3a66c8b52 req-d435ec17-d780-4ac8-966a-e1ba29e9d2ea service nova] Lock "382fa76b-b930-459e-b84e-da1b9eb74b74-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.122186] env[63372]: DEBUG oslo_concurrency.lockutils [req-e1037469-0611-4760-bcbf-caa3a66c8b52 req-d435ec17-d780-4ac8-966a-e1ba29e9d2ea service nova] Lock "382fa76b-b930-459e-b84e-da1b9eb74b74-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.122321] env[63372]: DEBUG nova.compute.manager [req-e1037469-0611-4760-bcbf-caa3a66c8b52 req-d435ec17-d780-4ac8-966a-e1ba29e9d2ea service nova] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] No waiting events found dispatching network-vif-plugged-777f211d-71d8-4f32-bfac-588b17d506f9 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 893.122495] env[63372]: WARNING nova.compute.manager [req-e1037469-0611-4760-bcbf-caa3a66c8b52 req-d435ec17-d780-4ac8-966a-e1ba29e9d2ea service nova] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Received unexpected event network-vif-plugged-777f211d-71d8-4f32-bfac-588b17d506f9 for instance with vm_state building and task_state spawning. [ 893.193911] env[63372]: DEBUG nova.compute.utils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 893.195708] env[63372]: DEBUG nova.compute.manager [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 893.195965] env[63372]: DEBUG nova.network.neutron [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 893.245653] env[63372]: DEBUG nova.policy [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '118d2f1de2be45cc8bb48bb75525d37e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'afe39cfedf214d50be775ef736f94da9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 893.273037] env[63372]: DEBUG nova.network.neutron [-] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.276942] env[63372]: DEBUG nova.network.neutron [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Successfully updated port: 777f211d-71d8-4f32-bfac-588b17d506f9 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 893.308164] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d09ba643-71d5-4651-8943-043344b8afb7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.328012] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233442b9-3398-405a-b3eb-26938a75aab2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.347761] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance '19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5' progress to 67 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 893.393393] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e15ad932-7f60-46fa-8233-54184a14e614 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.401151] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb72d432-7dd8-4377-842b-9e46ff463ec8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.426231] env[63372]: DEBUG nova.compute.manager [req-fa1f3085-eaf1-4bee-8b9a-83e8300f5170 req-5dd8373a-f2f3-41d7-a201-391f4714845e service nova] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Detach interface failed, port_id=75cfde99-66f3-4e6e-8755-061b37055d28, reason: Instance 1f52f9ee-2af3-4321-8ec6-5d5e841ed880 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 893.580671] env[63372]: DEBUG nova.network.neutron [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Successfully created port: 43013e6e-28ee-4f8c-afa8-20c12b01c0f8 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 893.669872] env[63372]: DEBUG nova.network.neutron [-] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.707069] env[63372]: DEBUG nova.compute.manager [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 893.779239] env[63372]: INFO nova.compute.manager [-] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Took 1.32 seconds to deallocate network for instance. [ 893.779769] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "refresh_cache-382fa76b-b930-459e-b84e-da1b9eb74b74" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.780046] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquired lock "refresh_cache-382fa76b-b930-459e-b84e-da1b9eb74b74" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.781294] env[63372]: DEBUG nova.network.neutron [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 893.927524] env[63372]: DEBUG nova.network.neutron [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Port e1486286-9147-4e5e-88c1-60893a3338dd binding to destination host cpu-1 is already ACTIVE {{(pid=63372) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 893.943281] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de63173b-b9e9-4e7f-bf97-63ee8e582b6e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.951212] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f39dd3-0ac5-4a1c-a867-d947b2aa7895 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.983733] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5bc0888-d16f-444b-9f92-f547d641191f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.995407] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3cdf6ab-2e17-41bd-addd-91865f1dabf6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.011117] env[63372]: DEBUG nova.compute.provider_tree [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.173184] env[63372]: INFO nova.compute.manager [-] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Took 1.24 seconds to deallocate network for instance. [ 894.287815] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.355457] env[63372]: DEBUG nova.network.neutron [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 894.513774] env[63372]: DEBUG nova.scheduler.client.report [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.623481] env[63372]: DEBUG nova.network.neutron [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Updating instance_info_cache with network_info: [{"id": "777f211d-71d8-4f32-bfac-588b17d506f9", "address": "fa:16:3e:88:19:9c", "network": {"id": "b924481a-4e51-4483-a710-f01fc39c851c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1289474667-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b6afb5579f2b48e4b094f6490136417f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap777f211d-71", "ovs_interfaceid": "777f211d-71d8-4f32-bfac-588b17d506f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.653950] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 894.654214] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 894.679490] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.715073] env[63372]: DEBUG nova.compute.manager [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 894.772400] env[63372]: DEBUG nova.virt.hardware [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 894.772686] env[63372]: DEBUG nova.virt.hardware [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 894.772837] env[63372]: DEBUG nova.virt.hardware [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.773032] env[63372]: DEBUG nova.virt.hardware [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 894.773190] env[63372]: DEBUG nova.virt.hardware [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.773339] env[63372]: DEBUG nova.virt.hardware [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 894.773548] env[63372]: DEBUG nova.virt.hardware [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 894.773710] env[63372]: DEBUG nova.virt.hardware [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 894.773877] env[63372]: DEBUG nova.virt.hardware [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 894.774054] env[63372]: DEBUG nova.virt.hardware [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 894.774231] env[63372]: DEBUG nova.virt.hardware [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 894.775258] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f901b5cb-79ad-45d8-9f94-5f05af4da45d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.783238] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a82c52-cf4d-4802-8241-42cad8583613 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.915597] env[63372]: DEBUG nova.compute.manager [req-d3738ef5-e829-4618-9ada-40b3705c09c9 req-b7e04b07-1517-4257-bbc7-57366a949499 service nova] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Received event network-vif-deleted-617ca099-5dd5-46c3-afc2-f80f4dbe242f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.948613] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.948613] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.948613] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.022718] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.023808] env[63372]: DEBUG nova.compute.manager [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 895.026423] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.778s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.031620] env[63372]: INFO nova.compute.claims [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.083691] env[63372]: DEBUG nova.compute.manager [req-45d69901-e9c1-4534-af01-9a1a503ff07f req-a736c0fd-9cce-4aee-a37d-9bb512d79995 service nova] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Received event network-vif-plugged-43013e6e-28ee-4f8c-afa8-20c12b01c0f8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.083912] env[63372]: DEBUG oslo_concurrency.lockutils [req-45d69901-e9c1-4534-af01-9a1a503ff07f req-a736c0fd-9cce-4aee-a37d-9bb512d79995 service nova] Acquiring lock "75560852-8c6d-4220-b251-81e8721a4b7b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.084134] env[63372]: DEBUG oslo_concurrency.lockutils [req-45d69901-e9c1-4534-af01-9a1a503ff07f req-a736c0fd-9cce-4aee-a37d-9bb512d79995 service nova] Lock "75560852-8c6d-4220-b251-81e8721a4b7b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.084303] env[63372]: DEBUG oslo_concurrency.lockutils [req-45d69901-e9c1-4534-af01-9a1a503ff07f req-a736c0fd-9cce-4aee-a37d-9bb512d79995 service nova] Lock "75560852-8c6d-4220-b251-81e8721a4b7b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.084469] env[63372]: DEBUG nova.compute.manager [req-45d69901-e9c1-4534-af01-9a1a503ff07f req-a736c0fd-9cce-4aee-a37d-9bb512d79995 service nova] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] No waiting events found dispatching network-vif-plugged-43013e6e-28ee-4f8c-afa8-20c12b01c0f8 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 895.084670] env[63372]: WARNING nova.compute.manager [req-45d69901-e9c1-4534-af01-9a1a503ff07f req-a736c0fd-9cce-4aee-a37d-9bb512d79995 service nova] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Received unexpected event network-vif-plugged-43013e6e-28ee-4f8c-afa8-20c12b01c0f8 for instance with vm_state building and task_state spawning. [ 895.127600] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Releasing lock "refresh_cache-382fa76b-b930-459e-b84e-da1b9eb74b74" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.127929] env[63372]: DEBUG nova.compute.manager [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Instance network_info: |[{"id": "777f211d-71d8-4f32-bfac-588b17d506f9", "address": "fa:16:3e:88:19:9c", "network": {"id": "b924481a-4e51-4483-a710-f01fc39c851c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1289474667-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b6afb5579f2b48e4b094f6490136417f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap777f211d-71", "ovs_interfaceid": "777f211d-71d8-4f32-bfac-588b17d506f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 895.128393] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:19:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '40c947c4-f471-4d48-8e43-fee54198107e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '777f211d-71d8-4f32-bfac-588b17d506f9', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 895.136250] env[63372]: DEBUG oslo.service.loopingcall [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.136473] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 895.136702] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce3c224f-d6b1-4a6c-a799-3098d52e194e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.156799] env[63372]: DEBUG nova.compute.manager [req-0aa7b097-8e89-4354-bbaa-c6922836efab req-77386f4b-df8a-4f4b-8c14-8f89b9b7925f service nova] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Received event network-changed-777f211d-71d8-4f32-bfac-588b17d506f9 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.156897] env[63372]: DEBUG nova.compute.manager [req-0aa7b097-8e89-4354-bbaa-c6922836efab req-77386f4b-df8a-4f4b-8c14-8f89b9b7925f service nova] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Refreshing instance network info cache due to event network-changed-777f211d-71d8-4f32-bfac-588b17d506f9. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 895.157125] env[63372]: DEBUG oslo_concurrency.lockutils [req-0aa7b097-8e89-4354-bbaa-c6922836efab req-77386f4b-df8a-4f4b-8c14-8f89b9b7925f service nova] Acquiring lock "refresh_cache-382fa76b-b930-459e-b84e-da1b9eb74b74" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.157303] env[63372]: DEBUG oslo_concurrency.lockutils [req-0aa7b097-8e89-4354-bbaa-c6922836efab req-77386f4b-df8a-4f4b-8c14-8f89b9b7925f service nova] Acquired lock "refresh_cache-382fa76b-b930-459e-b84e-da1b9eb74b74" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.157431] env[63372]: DEBUG nova.network.neutron [req-0aa7b097-8e89-4354-bbaa-c6922836efab req-77386f4b-df8a-4f4b-8c14-8f89b9b7925f service nova] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Refreshing network info cache for port 777f211d-71d8-4f32-bfac-588b17d506f9 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 895.163055] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 895.163055] env[63372]: value = "task-1024371" [ 895.163055] env[63372]: _type = "Task" [ 895.163055] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.164042] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.164042] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Starting heal instance info cache {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 895.174437] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024371, 'name': CreateVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.535723] env[63372]: DEBUG nova.compute.utils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 895.537076] env[63372]: DEBUG nova.compute.manager [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 895.537252] env[63372]: DEBUG nova.network.neutron [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 895.600732] env[63372]: DEBUG nova.network.neutron [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Successfully updated port: 43013e6e-28ee-4f8c-afa8-20c12b01c0f8 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 895.620250] env[63372]: DEBUG nova.policy [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb233d4b6fe54fefae3046d62d167e34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d0c0f315749429dbae22a72ae82b1b2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 895.676792] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024371, 'name': CreateVM_Task, 'duration_secs': 0.297057} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.676978] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 895.677654] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.677818] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.678158] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 895.678411] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e33d4e8-7dfd-45b0-9e9c-f00a9218fd7b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.682857] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 895.682857] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f0675b-b947-8bad-c03e-fa96f44660b6" [ 895.682857] env[63372]: _type = "Task" [ 895.682857] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.690991] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f0675b-b947-8bad-c03e-fa96f44660b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.872031] env[63372]: DEBUG nova.network.neutron [req-0aa7b097-8e89-4354-bbaa-c6922836efab req-77386f4b-df8a-4f4b-8c14-8f89b9b7925f service nova] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Updated VIF entry in instance network info cache for port 777f211d-71d8-4f32-bfac-588b17d506f9. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 895.872413] env[63372]: DEBUG nova.network.neutron [req-0aa7b097-8e89-4354-bbaa-c6922836efab req-77386f4b-df8a-4f4b-8c14-8f89b9b7925f service nova] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Updating instance_info_cache with network_info: [{"id": "777f211d-71d8-4f32-bfac-588b17d506f9", "address": "fa:16:3e:88:19:9c", "network": {"id": "b924481a-4e51-4483-a710-f01fc39c851c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1289474667-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b6afb5579f2b48e4b094f6490136417f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40c947c4-f471-4d48-8e43-fee54198107e", "external-id": "nsx-vlan-transportzone-203", "segmentation_id": 203, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap777f211d-71", "ovs_interfaceid": "777f211d-71d8-4f32-bfac-588b17d506f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.989014] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.989014] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.989014] env[63372]: DEBUG nova.network.neutron [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 896.040184] env[63372]: DEBUG nova.compute.manager [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 896.090083] env[63372]: DEBUG nova.network.neutron [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Successfully created port: 7d3f76e6-e43d-4af6-84df-522498d9968a {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 896.104377] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "refresh_cache-75560852-8c6d-4220-b251-81e8721a4b7b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.104526] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "refresh_cache-75560852-8c6d-4220-b251-81e8721a4b7b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.104674] env[63372]: DEBUG nova.network.neutron [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 896.193860] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f0675b-b947-8bad-c03e-fa96f44660b6, 'name': SearchDatastore_Task, 'duration_secs': 0.009559} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.197209] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.197459] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 896.197699] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.197846] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.198038] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.198504] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1042c9c9-6a32-4f6f-8ab1-3cdaa68f8123 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.207381] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.207565] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 896.210460] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d185417-e249-4ed0-9056-60298815a753 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.215808] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 896.215808] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5255ca48-3a02-b8d4-4185-9cb68efe0477" [ 896.215808] env[63372]: _type = "Task" [ 896.215808] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.223654] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5255ca48-3a02-b8d4-4185-9cb68efe0477, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.284969] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72bcfac4-c4c9-4dfc-9c44-628e1d91f8e8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.292345] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e7dedf-e80a-459d-991f-1f6feff106d5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.323267] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c080c661-c0f3-49e5-904f-420c9ac53825 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.331570] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac0a40d-110c-4524-b4a0-3a3653069b64 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.349477] env[63372]: DEBUG nova.compute.provider_tree [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.375499] env[63372]: DEBUG oslo_concurrency.lockutils [req-0aa7b097-8e89-4354-bbaa-c6922836efab req-77386f4b-df8a-4f4b-8c14-8f89b9b7925f service nova] Releasing lock "refresh_cache-382fa76b-b930-459e-b84e-da1b9eb74b74" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.647011] env[63372]: DEBUG nova.network.neutron [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 896.698442] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "refresh_cache-ac90a156-be00-4f62-a76e-e08914531167" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.698442] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquired lock "refresh_cache-ac90a156-be00-4f62-a76e-e08914531167" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.698577] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: ac90a156-be00-4f62-a76e-e08914531167] Forcefully refreshing network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 896.726271] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5255ca48-3a02-b8d4-4185-9cb68efe0477, 'name': SearchDatastore_Task, 'duration_secs': 0.010321} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.727071] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b88399f4-9670-4dc6-a226-3112910e6e8f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.729748] env[63372]: DEBUG nova.network.neutron [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance_info_cache with network_info: [{"id": "e1486286-9147-4e5e-88c1-60893a3338dd", "address": "fa:16:3e:6c:92:ac", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1486286-91", "ovs_interfaceid": "e1486286-9147-4e5e-88c1-60893a3338dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.733630] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 896.733630] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52597b38-ca26-3aad-6ab9-d6c16be662e3" [ 896.733630] env[63372]: _type = "Task" [ 896.733630] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.741180] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52597b38-ca26-3aad-6ab9-d6c16be662e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.790140] env[63372]: DEBUG nova.network.neutron [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Updating instance_info_cache with network_info: [{"id": "43013e6e-28ee-4f8c-afa8-20c12b01c0f8", "address": "fa:16:3e:78:f6:1b", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43013e6e-28", "ovs_interfaceid": "43013e6e-28ee-4f8c-afa8-20c12b01c0f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.856363] env[63372]: DEBUG nova.scheduler.client.report [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.999424] env[63372]: DEBUG nova.compute.manager [req-2c98bb6e-5356-4e32-a98b-5c5952549f24 req-91036e77-a13e-457e-9b35-f5808f89eca9 service nova] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Received event network-changed-43013e6e-28ee-4f8c-afa8-20c12b01c0f8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.999660] env[63372]: DEBUG nova.compute.manager [req-2c98bb6e-5356-4e32-a98b-5c5952549f24 req-91036e77-a13e-457e-9b35-f5808f89eca9 service nova] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Refreshing instance network info cache due to event network-changed-43013e6e-28ee-4f8c-afa8-20c12b01c0f8. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.999763] env[63372]: DEBUG oslo_concurrency.lockutils [req-2c98bb6e-5356-4e32-a98b-5c5952549f24 req-91036e77-a13e-457e-9b35-f5808f89eca9 service nova] Acquiring lock "refresh_cache-75560852-8c6d-4220-b251-81e8721a4b7b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.052067] env[63372]: DEBUG nova.compute.manager [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 897.108746] env[63372]: DEBUG nova.virt.hardware [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.109071] env[63372]: DEBUG nova.virt.hardware [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.109247] env[63372]: DEBUG nova.virt.hardware [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.109433] env[63372]: DEBUG nova.virt.hardware [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.109580] env[63372]: DEBUG nova.virt.hardware [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.109728] env[63372]: DEBUG nova.virt.hardware [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.109934] env[63372]: DEBUG nova.virt.hardware [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.110154] env[63372]: DEBUG nova.virt.hardware [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.110335] env[63372]: DEBUG nova.virt.hardware [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.110499] env[63372]: DEBUG nova.virt.hardware [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.110675] env[63372]: DEBUG nova.virt.hardware [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.111545] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7607ce-e036-484a-8122-15dcebb19ba0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.119231] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-801b0543-7b67-477d-8e28-57e2ca3af0f9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.217917] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: ac90a156-be00-4f62-a76e-e08914531167] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 897.231836] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.244099] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52597b38-ca26-3aad-6ab9-d6c16be662e3, 'name': SearchDatastore_Task, 'duration_secs': 0.009744} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.246223] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.246475] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 382fa76b-b930-459e-b84e-da1b9eb74b74/382fa76b-b930-459e-b84e-da1b9eb74b74.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 897.246921] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9cb4c4f8-b478-4e2d-82ad-944d185a5014 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.253796] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 897.253796] env[63372]: value = "task-1024372" [ 897.253796] env[63372]: _type = "Task" [ 897.253796] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.261602] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024372, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.292690] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "refresh_cache-75560852-8c6d-4220-b251-81e8721a4b7b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.292993] env[63372]: DEBUG nova.compute.manager [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Instance network_info: |[{"id": "43013e6e-28ee-4f8c-afa8-20c12b01c0f8", "address": "fa:16:3e:78:f6:1b", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43013e6e-28", "ovs_interfaceid": "43013e6e-28ee-4f8c-afa8-20c12b01c0f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 897.293310] env[63372]: DEBUG oslo_concurrency.lockutils [req-2c98bb6e-5356-4e32-a98b-5c5952549f24 req-91036e77-a13e-457e-9b35-f5808f89eca9 service nova] Acquired lock "refresh_cache-75560852-8c6d-4220-b251-81e8721a4b7b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.293492] env[63372]: DEBUG nova.network.neutron [req-2c98bb6e-5356-4e32-a98b-5c5952549f24 req-91036e77-a13e-457e-9b35-f5808f89eca9 service nova] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Refreshing network info cache for port 43013e6e-28ee-4f8c-afa8-20c12b01c0f8 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 897.294621] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:f6:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b21ab10-d886-4453-9472-9e11fb3c450d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43013e6e-28ee-4f8c-afa8-20c12b01c0f8', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.302283] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Creating folder: Project (afe39cfedf214d50be775ef736f94da9). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 897.303197] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-325ea90e-b590-433e-b76b-0c0b5bb5fe96 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.314211] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Created folder: Project (afe39cfedf214d50be775ef736f94da9) in parent group-v227230. [ 897.314841] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Creating folder: Instances. Parent ref: group-v227398. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 897.314841] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-16db1f6d-7650-4c10-8827-78e98c2a03cd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.324393] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Created folder: Instances in parent group-v227398. [ 897.324629] env[63372]: DEBUG oslo.service.loopingcall [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.324815] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 897.325032] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-800f7a08-2116-454d-9bc3-47707791abe8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.344948] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.344948] env[63372]: value = "task-1024375" [ 897.344948] env[63372]: _type = "Task" [ 897.344948] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.352824] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024375, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.360936] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.334s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.361450] env[63372]: DEBUG nova.compute.manager [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 897.363784] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.799s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.365214] env[63372]: INFO nova.compute.claims [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 897.763864] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024372, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46514} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.764190] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 382fa76b-b930-459e-b84e-da1b9eb74b74/382fa76b-b930-459e-b84e-da1b9eb74b74.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 897.764426] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 897.764657] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-188430a8-448d-4f85-9a9c-c9ee347586eb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.770704] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 897.770704] env[63372]: value = "task-1024376" [ 897.770704] env[63372]: _type = "Task" [ 897.770704] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.780716] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024376, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.782374] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc85a72-c7ed-45d2-bbae-79c5fa1d92ce {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.809590] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a61eca-15f3-4cce-836b-87c82a8254a6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.816562] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance '19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5' progress to 83 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 897.823227] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.857908] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024375, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.869293] env[63372]: DEBUG nova.compute.utils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 897.873259] env[63372]: DEBUG nova.compute.manager [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 897.873436] env[63372]: DEBUG nova.network.neutron [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 897.888439] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: ac90a156-be00-4f62-a76e-e08914531167] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.952283] env[63372]: DEBUG nova.policy [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d822682f06640489a48f8577acf4389', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e46b6acae4a4b71bc390bd513ddd2f1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 898.068459] env[63372]: DEBUG nova.compute.manager [req-9e2111f0-86c2-4210-b355-5bd3af3621cc req-cf8bed81-e789-466b-8eff-38bb413e0ec0 service nova] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Received event network-vif-plugged-7d3f76e6-e43d-4af6-84df-522498d9968a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.068844] env[63372]: DEBUG oslo_concurrency.lockutils [req-9e2111f0-86c2-4210-b355-5bd3af3621cc req-cf8bed81-e789-466b-8eff-38bb413e0ec0 service nova] Acquiring lock "ba1daa0a-dcf8-4586-9964-a962fcc94196-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.069098] env[63372]: DEBUG oslo_concurrency.lockutils [req-9e2111f0-86c2-4210-b355-5bd3af3621cc req-cf8bed81-e789-466b-8eff-38bb413e0ec0 service nova] Lock "ba1daa0a-dcf8-4586-9964-a962fcc94196-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.069309] env[63372]: DEBUG oslo_concurrency.lockutils [req-9e2111f0-86c2-4210-b355-5bd3af3621cc req-cf8bed81-e789-466b-8eff-38bb413e0ec0 service nova] Lock "ba1daa0a-dcf8-4586-9964-a962fcc94196-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.069490] env[63372]: DEBUG nova.compute.manager [req-9e2111f0-86c2-4210-b355-5bd3af3621cc req-cf8bed81-e789-466b-8eff-38bb413e0ec0 service nova] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] No waiting events found dispatching network-vif-plugged-7d3f76e6-e43d-4af6-84df-522498d9968a {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 898.069672] env[63372]: WARNING nova.compute.manager [req-9e2111f0-86c2-4210-b355-5bd3af3621cc req-cf8bed81-e789-466b-8eff-38bb413e0ec0 service nova] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Received unexpected event network-vif-plugged-7d3f76e6-e43d-4af6-84df-522498d9968a for instance with vm_state building and task_state spawning. [ 898.086977] env[63372]: DEBUG nova.network.neutron [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Successfully updated port: 7d3f76e6-e43d-4af6-84df-522498d9968a {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 898.121755] env[63372]: DEBUG nova.network.neutron [req-2c98bb6e-5356-4e32-a98b-5c5952549f24 req-91036e77-a13e-457e-9b35-f5808f89eca9 service nova] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Updated VIF entry in instance network info cache for port 43013e6e-28ee-4f8c-afa8-20c12b01c0f8. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 898.121755] env[63372]: DEBUG nova.network.neutron [req-2c98bb6e-5356-4e32-a98b-5c5952549f24 req-91036e77-a13e-457e-9b35-f5808f89eca9 service nova] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Updating instance_info_cache with network_info: [{"id": "43013e6e-28ee-4f8c-afa8-20c12b01c0f8", "address": "fa:16:3e:78:f6:1b", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43013e6e-28", "ovs_interfaceid": "43013e6e-28ee-4f8c-afa8-20c12b01c0f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.279742] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024376, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066423} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.283529] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.283529] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d3d399-e019-402d-8faf-fb412e672c64 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.304542] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 382fa76b-b930-459e-b84e-da1b9eb74b74/382fa76b-b930-459e-b84e-da1b9eb74b74.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.305445] env[63372]: DEBUG nova.network.neutron [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Successfully created port: 2454845a-285a-4d06-8f95-123e004fae02 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 898.307412] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-613517c3-f791-44a8-9e35-0d8b63912ac9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.326226] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f07aea66-8589-496f-98ff-b4262292b006 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance '19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5' progress to 100 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 898.331372] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 898.331372] env[63372]: value = "task-1024377" [ 898.331372] env[63372]: _type = "Task" [ 898.331372] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.339355] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024377, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.356703] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024375, 'name': CreateVM_Task, 'duration_secs': 0.562616} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.358081] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 898.358081] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.358081] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.358870] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 898.358870] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-809756b2-3302-4062-9a63-37eef25833c9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.362810] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 898.362810] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52859c00-0ae4-972a-942b-b836f63af913" [ 898.362810] env[63372]: _type = "Task" [ 898.362810] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.370357] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52859c00-0ae4-972a-942b-b836f63af913, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.378349] env[63372]: DEBUG nova.compute.manager [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 898.380475] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5387a3b-2d13-4b4c-85fc-12609e8e8234 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "interface-13bde897-8446-42a2-b02d-2f5b48e6f432-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.380698] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5387a3b-2d13-4b4c-85fc-12609e8e8234 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-13bde897-8446-42a2-b02d-2f5b48e6f432-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.381012] env[63372]: DEBUG nova.objects.instance [None req-e5387a3b-2d13-4b4c-85fc-12609e8e8234 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'flavor' on Instance uuid 13bde897-8446-42a2-b02d-2f5b48e6f432 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.391038] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Releasing lock "refresh_cache-ac90a156-be00-4f62-a76e-e08914531167" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.394990] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: ac90a156-be00-4f62-a76e-e08914531167] Updated the network info_cache for instance {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 898.394990] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 898.394990] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 898.394990] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 898.394990] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 898.394990] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 898.394990] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 898.394990] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63372) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 898.394990] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager.update_available_resource {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 898.589034] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "refresh_cache-ba1daa0a-dcf8-4586-9964-a962fcc94196" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.589231] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "refresh_cache-ba1daa0a-dcf8-4586-9964-a962fcc94196" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.589562] env[63372]: DEBUG nova.network.neutron [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 898.624114] env[63372]: DEBUG oslo_concurrency.lockutils [req-2c98bb6e-5356-4e32-a98b-5c5952549f24 req-91036e77-a13e-457e-9b35-f5808f89eca9 service nova] Releasing lock "refresh_cache-75560852-8c6d-4220-b251-81e8721a4b7b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.633358] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46642857-1605-4f1c-9ef7-2f4dcbd443d1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.641545] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9523a462-9537-4268-8e68-4840cfe1a96e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.671888] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f6cf34-f5ba-4399-83cb-43c56657f45e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.679150] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7adc75f-ff90-4405-a65c-000d259d783a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.691982] env[63372]: DEBUG nova.compute.provider_tree [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 898.846515] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024377, 'name': ReconfigVM_Task, 'duration_secs': 0.276459} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.846515] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 382fa76b-b930-459e-b84e-da1b9eb74b74/382fa76b-b930-459e-b84e-da1b9eb74b74.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.847023] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e09a3792-ee25-4e3d-a92f-f42264f89f56 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.853031] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 898.853031] env[63372]: value = "task-1024378" [ 898.853031] env[63372]: _type = "Task" [ 898.853031] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.861721] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024378, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.872275] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52859c00-0ae4-972a-942b-b836f63af913, 'name': SearchDatastore_Task, 'duration_secs': 0.008502} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.872573] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.872811] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.873057] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.873215] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.873398] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.873650] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b817350-f02f-4690-9892-78f300f0461f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.886670] env[63372]: DEBUG nova.objects.instance [None req-e5387a3b-2d13-4b4c-85fc-12609e8e8234 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'pci_requests' on Instance uuid 13bde897-8446-42a2-b02d-2f5b48e6f432 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.887701] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.887836] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 898.889220] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b24b2de-ef97-4724-b2c4-98f420fe30cb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.895996] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.897900] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 898.897900] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52bbc376-c0fa-074d-650d-07dfe0cb731e" [ 898.897900] env[63372]: _type = "Task" [ 898.897900] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.905778] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52bbc376-c0fa-074d-650d-07dfe0cb731e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.123600] env[63372]: DEBUG nova.network.neutron [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 899.227820] env[63372]: DEBUG nova.scheduler.client.report [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 108 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 899.228079] env[63372]: DEBUG nova.compute.provider_tree [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 108 to 109 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 899.228264] env[63372]: DEBUG nova.compute.provider_tree [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 899.265503] env[63372]: DEBUG nova.network.neutron [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Updating instance_info_cache with network_info: [{"id": "7d3f76e6-e43d-4af6-84df-522498d9968a", "address": "fa:16:3e:70:db:3c", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d3f76e6-e4", "ovs_interfaceid": "7d3f76e6-e43d-4af6-84df-522498d9968a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.362713] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024378, 'name': Rename_Task, 'duration_secs': 0.146948} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.363017] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 899.363284] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-58beef27-ecc8-4bfe-8fe4-5d0a9dd2805e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.370494] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 899.370494] env[63372]: value = "task-1024379" [ 899.370494] env[63372]: _type = "Task" [ 899.370494] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.378525] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024379, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.390247] env[63372]: DEBUG nova.objects.base [None req-e5387a3b-2d13-4b4c-85fc-12609e8e8234 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Object Instance<13bde897-8446-42a2-b02d-2f5b48e6f432> lazy-loaded attributes: flavor,pci_requests {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 899.390798] env[63372]: DEBUG nova.network.neutron [None req-e5387a3b-2d13-4b4c-85fc-12609e8e8234 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 899.393037] env[63372]: DEBUG nova.compute.manager [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 899.408300] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52bbc376-c0fa-074d-650d-07dfe0cb731e, 'name': SearchDatastore_Task, 'duration_secs': 0.00789} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.409130] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67561456-dfcb-4d7c-8a5f-6d62ae399bfb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.414766] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 899.414766] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]526933f7-1082-955b-03b4-283c93238c96" [ 899.414766] env[63372]: _type = "Task" [ 899.414766] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.424560] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526933f7-1082-955b-03b4-283c93238c96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.438714] env[63372]: DEBUG nova.virt.hardware [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.438714] env[63372]: DEBUG nova.virt.hardware [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.438714] env[63372]: DEBUG nova.virt.hardware [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.438714] env[63372]: DEBUG nova.virt.hardware [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.438983] env[63372]: DEBUG nova.virt.hardware [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.438983] env[63372]: DEBUG nova.virt.hardware [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.440079] env[63372]: DEBUG nova.virt.hardware [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.440079] env[63372]: DEBUG nova.virt.hardware [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.440079] env[63372]: DEBUG nova.virt.hardware [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.440079] env[63372]: DEBUG nova.virt.hardware [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.440079] env[63372]: DEBUG nova.virt.hardware [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.440953] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab28380-0953-4dfe-aa0c-e55ffdcf5b08 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.450500] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65335e31-7af5-44d3-8383-716eb753d56b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.488523] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e5387a3b-2d13-4b4c-85fc-12609e8e8234 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-13bde897-8446-42a2-b02d-2f5b48e6f432-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.108s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.737083] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.737694] env[63372]: DEBUG nova.compute.manager [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 899.740708] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.129s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.740923] env[63372]: DEBUG nova.objects.instance [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lazy-loading 'resources' on Instance uuid 99f901a6-9bb3-4403-af0c-c8900f655cb3 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.768256] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "refresh_cache-ba1daa0a-dcf8-4586-9964-a962fcc94196" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.768784] env[63372]: DEBUG nova.compute.manager [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Instance network_info: |[{"id": "7d3f76e6-e43d-4af6-84df-522498d9968a", "address": "fa:16:3e:70:db:3c", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d3f76e6-e4", "ovs_interfaceid": "7d3f76e6-e43d-4af6-84df-522498d9968a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 899.769351] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:db:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60badc2d-69d2-467d-a92e-98511f5cb0b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d3f76e6-e43d-4af6-84df-522498d9968a', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 899.777691] env[63372]: DEBUG oslo.service.loopingcall [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.778657] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 899.778905] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ede466b-f44b-4f8d-b4dd-82ce6f5ebca5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.799311] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 899.799311] env[63372]: value = "task-1024380" [ 899.799311] env[63372]: _type = "Task" [ 899.799311] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.809065] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024380, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.880404] env[63372]: DEBUG oslo_vmware.api [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024379, 'name': PowerOnVM_Task, 'duration_secs': 0.427041} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.881896] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 899.882127] env[63372]: INFO nova.compute.manager [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Took 7.54 seconds to spawn the instance on the hypervisor. [ 899.882319] env[63372]: DEBUG nova.compute.manager [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.883401] env[63372]: DEBUG nova.compute.manager [req-5d03744c-7bc2-4e06-a061-c52134f83518 req-16851dd6-c86c-4a72-af62-f28b75971485 service nova] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Received event network-vif-plugged-2454845a-285a-4d06-8f95-123e004fae02 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.883598] env[63372]: DEBUG oslo_concurrency.lockutils [req-5d03744c-7bc2-4e06-a061-c52134f83518 req-16851dd6-c86c-4a72-af62-f28b75971485 service nova] Acquiring lock "d568b727-7836-4b1d-9f27-f159227e46f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.883978] env[63372]: DEBUG oslo_concurrency.lockutils [req-5d03744c-7bc2-4e06-a061-c52134f83518 req-16851dd6-c86c-4a72-af62-f28b75971485 service nova] Lock "d568b727-7836-4b1d-9f27-f159227e46f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.883978] env[63372]: DEBUG oslo_concurrency.lockutils [req-5d03744c-7bc2-4e06-a061-c52134f83518 req-16851dd6-c86c-4a72-af62-f28b75971485 service nova] Lock "d568b727-7836-4b1d-9f27-f159227e46f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.884133] env[63372]: DEBUG nova.compute.manager [req-5d03744c-7bc2-4e06-a061-c52134f83518 req-16851dd6-c86c-4a72-af62-f28b75971485 service nova] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] No waiting events found dispatching network-vif-plugged-2454845a-285a-4d06-8f95-123e004fae02 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 899.884293] env[63372]: WARNING nova.compute.manager [req-5d03744c-7bc2-4e06-a061-c52134f83518 req-16851dd6-c86c-4a72-af62-f28b75971485 service nova] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Received unexpected event network-vif-plugged-2454845a-285a-4d06-8f95-123e004fae02 for instance with vm_state building and task_state spawning. [ 899.885216] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84a6558-3024-491f-b837-330229bdfa25 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.895982] env[63372]: DEBUG nova.network.neutron [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Successfully updated port: 2454845a-285a-4d06-8f95-123e004fae02 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 899.925086] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526933f7-1082-955b-03b4-283c93238c96, 'name': SearchDatastore_Task, 'duration_secs': 0.009279} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.925217] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.925827] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 75560852-8c6d-4220-b251-81e8721a4b7b/75560852-8c6d-4220-b251-81e8721a4b7b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 899.925827] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a9ced5af-33fb-421f-b7f7-6e03de3c8dba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.932259] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 899.932259] env[63372]: value = "task-1024381" [ 899.932259] env[63372]: _type = "Task" [ 899.932259] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.940773] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024381, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.196829] env[63372]: DEBUG nova.compute.manager [req-d380b42b-45ac-492d-8666-672e1ae8fa09 req-c4c2d40c-7e9a-4e7e-a052-bf476d4c1417 service nova] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Received event network-changed-7d3f76e6-e43d-4af6-84df-522498d9968a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.197174] env[63372]: DEBUG nova.compute.manager [req-d380b42b-45ac-492d-8666-672e1ae8fa09 req-c4c2d40c-7e9a-4e7e-a052-bf476d4c1417 service nova] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Refreshing instance network info cache due to event network-changed-7d3f76e6-e43d-4af6-84df-522498d9968a. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 900.197365] env[63372]: DEBUG oslo_concurrency.lockutils [req-d380b42b-45ac-492d-8666-672e1ae8fa09 req-c4c2d40c-7e9a-4e7e-a052-bf476d4c1417 service nova] Acquiring lock "refresh_cache-ba1daa0a-dcf8-4586-9964-a962fcc94196" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.197478] env[63372]: DEBUG oslo_concurrency.lockutils [req-d380b42b-45ac-492d-8666-672e1ae8fa09 req-c4c2d40c-7e9a-4e7e-a052-bf476d4c1417 service nova] Acquired lock "refresh_cache-ba1daa0a-dcf8-4586-9964-a962fcc94196" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.197611] env[63372]: DEBUG nova.network.neutron [req-d380b42b-45ac-492d-8666-672e1ae8fa09 req-c4c2d40c-7e9a-4e7e-a052-bf476d4c1417 service nova] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Refreshing network info cache for port 7d3f76e6-e43d-4af6-84df-522498d9968a {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 900.248168] env[63372]: DEBUG nova.compute.utils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 900.250027] env[63372]: DEBUG nova.compute.manager [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 900.250027] env[63372]: DEBUG nova.network.neutron [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 900.290851] env[63372]: DEBUG nova.policy [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2253a9a409d4f1ba9176485655c0988', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dac773c2536745e285181426ae34bb96', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 900.309342] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024380, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.404759] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-d568b727-7836-4b1d-9f27-f159227e46f7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.404759] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-d568b727-7836-4b1d-9f27-f159227e46f7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.404759] env[63372]: DEBUG nova.network.neutron [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 900.410035] env[63372]: INFO nova.compute.manager [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Took 35.22 seconds to build instance. [ 900.445902] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024381, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.500067] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f9094a-2b92-40eb-bad2-b83a583cc336 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.507749] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef02a76-9000-4b66-8a52-4c43e1af0656 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.538692] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c955eb-01fe-4422-b798-b32bc2fd97bb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.547048] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e997b34-f0cf-40cd-b9b4-be5331f4ec47 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.560309] env[63372]: DEBUG nova.compute.provider_tree [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.695734] env[63372]: DEBUG nova.network.neutron [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Successfully created port: 8321625e-abd2-4cc3-b545-2190d9759158 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 900.752493] env[63372]: DEBUG nova.compute.manager [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 900.809617] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024380, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.913086] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d466ee0c-c72e-4133-8326-13e32eaec416 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "382fa76b-b930-459e-b84e-da1b9eb74b74" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.732s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.942333] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024381, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.525292} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.943250] env[63372]: DEBUG nova.network.neutron [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 900.945229] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 75560852-8c6d-4220-b251-81e8721a4b7b/75560852-8c6d-4220-b251-81e8721a4b7b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 900.945450] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.945712] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e514a9e4-b249-440a-9449-7efc67938fd1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.953156] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 900.953156] env[63372]: value = "task-1024382" [ 900.953156] env[63372]: _type = "Task" [ 900.953156] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.962196] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024382, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.065956] env[63372]: DEBUG nova.scheduler.client.report [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.071625] env[63372]: DEBUG nova.network.neutron [req-d380b42b-45ac-492d-8666-672e1ae8fa09 req-c4c2d40c-7e9a-4e7e-a052-bf476d4c1417 service nova] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Updated VIF entry in instance network info cache for port 7d3f76e6-e43d-4af6-84df-522498d9968a. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 901.071625] env[63372]: DEBUG nova.network.neutron [req-d380b42b-45ac-492d-8666-672e1ae8fa09 req-c4c2d40c-7e9a-4e7e-a052-bf476d4c1417 service nova] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Updating instance_info_cache with network_info: [{"id": "7d3f76e6-e43d-4af6-84df-522498d9968a", "address": "fa:16:3e:70:db:3c", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d3f76e6-e4", "ovs_interfaceid": "7d3f76e6-e43d-4af6-84df-522498d9968a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.211152] env[63372]: DEBUG nova.network.neutron [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Updating instance_info_cache with network_info: [{"id": "2454845a-285a-4d06-8f95-123e004fae02", "address": "fa:16:3e:83:4b:f8", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2454845a-28", "ovs_interfaceid": "2454845a-285a-4d06-8f95-123e004fae02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.310039] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024380, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.314701] env[63372]: DEBUG oslo_concurrency.lockutils [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.315521] env[63372]: DEBUG oslo_concurrency.lockutils [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.315521] env[63372]: DEBUG nova.compute.manager [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Going to confirm migration 2 {{(pid=63372) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 901.463412] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024382, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058592} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.463701] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.464477] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb6dda8-5c3b-47c9-ae8a-c320a95978f6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.486106] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 75560852-8c6d-4220-b251-81e8721a4b7b/75560852-8c6d-4220-b251-81e8721a4b7b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.486675] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9135dcc-bcef-4605-a930-f39bb3ddcfc9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.506130] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 901.506130] env[63372]: value = "task-1024383" [ 901.506130] env[63372]: _type = "Task" [ 901.506130] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.515180] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024383, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.571402] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.831s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.573828] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.051s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.574101] env[63372]: DEBUG nova.objects.instance [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lazy-loading 'resources' on Instance uuid ac90a156-be00-4f62-a76e-e08914531167 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.577057] env[63372]: DEBUG oslo_concurrency.lockutils [req-d380b42b-45ac-492d-8666-672e1ae8fa09 req-c4c2d40c-7e9a-4e7e-a052-bf476d4c1417 service nova] Releasing lock "refresh_cache-ba1daa0a-dcf8-4586-9964-a962fcc94196" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.632664] env[63372]: INFO nova.scheduler.client.report [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Deleted allocations for instance 99f901a6-9bb3-4403-af0c-c8900f655cb3 [ 901.670435] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "interface-13bde897-8446-42a2-b02d-2f5b48e6f432-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.670747] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-13bde897-8446-42a2-b02d-2f5b48e6f432-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.671053] env[63372]: DEBUG nova.objects.instance [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'flavor' on Instance uuid 13bde897-8446-42a2-b02d-2f5b48e6f432 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.714482] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-d568b727-7836-4b1d-9f27-f159227e46f7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.714963] env[63372]: DEBUG nova.compute.manager [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Instance network_info: |[{"id": "2454845a-285a-4d06-8f95-123e004fae02", "address": "fa:16:3e:83:4b:f8", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2454845a-28", "ovs_interfaceid": "2454845a-285a-4d06-8f95-123e004fae02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 901.715532] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:4b:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d5d0e0d-cdec-474a-a891-a9ceff15a8b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2454845a-285a-4d06-8f95-123e004fae02', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.724697] env[63372]: DEBUG oslo.service.loopingcall [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.725751] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 901.725995] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1cbb1559-8672-4512-85a7-86bac5786558 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.745829] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.745829] env[63372]: value = "task-1024384" [ 901.745829] env[63372]: _type = "Task" [ 901.745829] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.755161] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024384, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.761547] env[63372]: DEBUG nova.compute.manager [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 901.813548] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024380, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.819788] env[63372]: DEBUG nova.virt.hardware [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.820141] env[63372]: DEBUG nova.virt.hardware [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.820371] env[63372]: DEBUG nova.virt.hardware [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.820637] env[63372]: DEBUG nova.virt.hardware [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.820850] env[63372]: DEBUG nova.virt.hardware [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.821077] env[63372]: DEBUG nova.virt.hardware [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.821370] env[63372]: DEBUG nova.virt.hardware [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.821600] env[63372]: DEBUG nova.virt.hardware [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.821836] env[63372]: DEBUG nova.virt.hardware [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.822097] env[63372]: DEBUG nova.virt.hardware [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.822355] env[63372]: DEBUG nova.virt.hardware [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.826859] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb030d0-c4e6-43a8-af8b-d7822fff26c6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.832932] env[63372]: DEBUG oslo_concurrency.lockutils [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "382fa76b-b930-459e-b84e-da1b9eb74b74" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.833251] env[63372]: DEBUG oslo_concurrency.lockutils [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "382fa76b-b930-459e-b84e-da1b9eb74b74" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.833535] env[63372]: DEBUG oslo_concurrency.lockutils [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "382fa76b-b930-459e-b84e-da1b9eb74b74-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.833799] env[63372]: DEBUG oslo_concurrency.lockutils [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "382fa76b-b930-459e-b84e-da1b9eb74b74-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.834058] env[63372]: DEBUG oslo_concurrency.lockutils [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "382fa76b-b930-459e-b84e-da1b9eb74b74-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.836485] env[63372]: INFO nova.compute.manager [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Terminating instance [ 901.839691] env[63372]: DEBUG nova.compute.manager [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 901.839984] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 901.841085] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba4eba3-adf2-4222-8efd-caf6119cf2b0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.847471] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffcdd09f-005f-40be-8e03-a43d7264476b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.853593] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 901.854145] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d272dee5-a617-4b3c-9725-c5464ff036f5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.864317] env[63372]: DEBUG oslo_vmware.api [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 901.864317] env[63372]: value = "task-1024385" [ 901.864317] env[63372]: _type = "Task" [ 901.864317] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.871442] env[63372]: DEBUG oslo_vmware.api [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024385, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.879795] env[63372]: DEBUG oslo_concurrency.lockutils [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.879974] env[63372]: DEBUG oslo_concurrency.lockutils [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.880214] env[63372]: DEBUG nova.network.neutron [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 901.880446] env[63372]: DEBUG nova.objects.instance [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'info_cache' on Instance uuid 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.017335] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024383, 'name': ReconfigVM_Task, 'duration_secs': 0.273477} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.017631] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 75560852-8c6d-4220-b251-81e8721a4b7b/75560852-8c6d-4220-b251-81e8721a4b7b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.018302] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2a089b6c-0a9b-4770-9b4f-26d54fed89ce {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.024386] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 902.024386] env[63372]: value = "task-1024386" [ 902.024386] env[63372]: _type = "Task" [ 902.024386] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.032752] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024386, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.140914] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b9886ec5-1435-4fc0-8e29-28bf4dea63d4 tempest-ListServerFiltersTestJSON-268662238 tempest-ListServerFiltersTestJSON-268662238-project-member] Lock "99f901a6-9bb3-4403-af0c-c8900f655cb3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.534s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.198337] env[63372]: DEBUG nova.compute.manager [req-1f7b2269-3bd1-40cc-8a9c-fcb8f33f714b req-91542834-d231-4e16-80d7-c6e7d3eaf739 service nova] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Received event network-changed-2454845a-285a-4d06-8f95-123e004fae02 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.198459] env[63372]: DEBUG nova.compute.manager [req-1f7b2269-3bd1-40cc-8a9c-fcb8f33f714b req-91542834-d231-4e16-80d7-c6e7d3eaf739 service nova] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Refreshing instance network info cache due to event network-changed-2454845a-285a-4d06-8f95-123e004fae02. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 902.198736] env[63372]: DEBUG oslo_concurrency.lockutils [req-1f7b2269-3bd1-40cc-8a9c-fcb8f33f714b req-91542834-d231-4e16-80d7-c6e7d3eaf739 service nova] Acquiring lock "refresh_cache-d568b727-7836-4b1d-9f27-f159227e46f7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.199070] env[63372]: DEBUG oslo_concurrency.lockutils [req-1f7b2269-3bd1-40cc-8a9c-fcb8f33f714b req-91542834-d231-4e16-80d7-c6e7d3eaf739 service nova] Acquired lock "refresh_cache-d568b727-7836-4b1d-9f27-f159227e46f7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.199070] env[63372]: DEBUG nova.network.neutron [req-1f7b2269-3bd1-40cc-8a9c-fcb8f33f714b req-91542834-d231-4e16-80d7-c6e7d3eaf739 service nova] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Refreshing network info cache for port 2454845a-285a-4d06-8f95-123e004fae02 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 902.257479] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024384, 'name': CreateVM_Task, 'duration_secs': 0.471198} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.260180] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 902.260410] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.260533] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.260875] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 902.261562] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d849692-946d-419d-8db6-23cf30e29478 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.264714] env[63372]: DEBUG nova.objects.instance [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'pci_requests' on Instance uuid 13bde897-8446-42a2-b02d-2f5b48e6f432 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.267176] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 902.267176] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529607ad-c09e-a277-0bfc-261f3130c8bb" [ 902.267176] env[63372]: _type = "Task" [ 902.267176] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.288077] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529607ad-c09e-a277-0bfc-261f3130c8bb, 'name': SearchDatastore_Task, 'duration_secs': 0.009548} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.288077] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.288077] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.288285] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.288429] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.288602] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.288849] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3eb9e49-2cda-485b-bd5d-959525c23d0e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.296945] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.297245] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 902.298934] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05aa7397-23ec-45d8-9a03-b1597c5d0127 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.306503] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 902.306503] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52ece01b-8d5c-0415-9b41-c1b457a38d93" [ 902.306503] env[63372]: _type = "Task" [ 902.306503] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.307288] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f8890a-0a09-4466-8a9b-82073d0f681e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.319541] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024380, 'name': CreateVM_Task, 'duration_secs': 2.235538} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.322774] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 902.324158] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.324405] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.324924] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 902.326385] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f131b6-1f95-4e72-a8a0-c1c6119edff0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.333492] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd7a1c2d-6aa3-4810-8092-de7f562573e4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.335221] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ece01b-8d5c-0415-9b41-c1b457a38d93, 'name': SearchDatastore_Task, 'duration_secs': 0.007643} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.336728] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-755ea40b-8503-4622-9a72-f0601059d6f2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.370736] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 902.370736] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5237add7-bb64-2bf2-e76f-471efd29f8d1" [ 902.370736] env[63372]: _type = "Task" [ 902.370736] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.375425] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f706f95-a25b-46d7-b89e-ab263fcc0a47 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.379865] env[63372]: DEBUG nova.network.neutron [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Successfully updated port: 8321625e-abd2-4cc3-b545-2190d9759158 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 902.385223] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 902.385223] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52af60cc-2f4a-6ae4-82a0-44cd57a21dc5" [ 902.385223] env[63372]: _type = "Task" [ 902.385223] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.396949] env[63372]: DEBUG oslo_vmware.api [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024385, 'name': PowerOffVM_Task, 'duration_secs': 0.218322} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.404047] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0bf3dc3-4750-4773-b784-42d31123ca2e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.406012] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 902.406216] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 902.406522] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5237add7-bb64-2bf2-e76f-471efd29f8d1, 'name': SearchDatastore_Task, 'duration_secs': 0.007828} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.409915] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-937330f3-30ba-4956-a7d3-d5cb437865a2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.409915] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.409915] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.409915] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.410378] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.410378] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.414471] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c6b76678-df3d-4c2a-98d8-5aadd0578908 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.416345] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52af60cc-2f4a-6ae4-82a0-44cd57a21dc5, 'name': SearchDatastore_Task, 'duration_secs': 0.008634} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.424481] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.424742] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] d568b727-7836-4b1d-9f27-f159227e46f7/d568b727-7836-4b1d-9f27-f159227e46f7.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 902.425204] env[63372]: DEBUG nova.compute.provider_tree [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.430020] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ee501f2-0c1f-425a-b707-0af1c676cb46 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.433269] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 902.433269] env[63372]: value = "task-1024388" [ 902.433269] env[63372]: _type = "Task" [ 902.433269] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.433516] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.433700] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 902.435038] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74fac074-22f4-4001-add3-a1166c20ca96 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.445558] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 902.445558] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5231cfc1-cc77-f455-43fc-bb172857e552" [ 902.445558] env[63372]: _type = "Task" [ 902.445558] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.448228] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024388, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.458628] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5231cfc1-cc77-f455-43fc-bb172857e552, 'name': SearchDatastore_Task, 'duration_secs': 0.007887} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.458628] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b07f757c-7000-40fe-890f-68f83830be71 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.461876] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 902.461876] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52929c9a-2260-450d-52ed-69f5aca071f2" [ 902.461876] env[63372]: _type = "Task" [ 902.461876] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.469330] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52929c9a-2260-450d-52ed-69f5aca071f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.492006] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 902.492419] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 902.492740] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Deleting the datastore file [datastore1] 382fa76b-b930-459e-b84e-da1b9eb74b74 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.493139] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-77fe8732-526c-44b4-89e2-d50d13316098 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.500759] env[63372]: DEBUG oslo_vmware.api [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for the task: (returnval){ [ 902.500759] env[63372]: value = "task-1024389" [ 902.500759] env[63372]: _type = "Task" [ 902.500759] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.511966] env[63372]: DEBUG oslo_vmware.api [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024389, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.534821] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024386, 'name': Rename_Task, 'duration_secs': 0.158401} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.534821] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 902.534821] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa94a933-df81-4348-baba-156bb39a16e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.540981] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 902.540981] env[63372]: value = "task-1024390" [ 902.540981] env[63372]: _type = "Task" [ 902.540981] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.549494] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024390, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.769344] env[63372]: DEBUG nova.objects.base [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Object Instance<13bde897-8446-42a2-b02d-2f5b48e6f432> lazy-loaded attributes: flavor,pci_requests {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 902.769629] env[63372]: DEBUG nova.network.neutron [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 902.868937] env[63372]: DEBUG nova.policy [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7ffe6776b664c00b9086046f9171ec6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46d2cb188a004dc385f2aa15458be929', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 902.887676] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "refresh_cache-326a9155-1c00-4c18-a9a7-6124e709c20b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.887825] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "refresh_cache-326a9155-1c00-4c18-a9a7-6124e709c20b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.887972] env[63372]: DEBUG nova.network.neutron [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 902.930143] env[63372]: DEBUG nova.scheduler.client.report [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 902.944531] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024388, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451165} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.944873] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] d568b727-7836-4b1d-9f27-f159227e46f7/d568b727-7836-4b1d-9f27-f159227e46f7.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 902.945154] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 902.945444] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-13af34c2-b2ad-4a86-91c3-afa90807e072 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.953644] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 902.953644] env[63372]: value = "task-1024391" [ 902.953644] env[63372]: _type = "Task" [ 902.953644] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.962639] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024391, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.972687] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52929c9a-2260-450d-52ed-69f5aca071f2, 'name': SearchDatastore_Task, 'duration_secs': 0.007535} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.972902] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.973229] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] ba1daa0a-dcf8-4586-9964-a962fcc94196/ba1daa0a-dcf8-4586-9964-a962fcc94196.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 902.973468] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0a6d4c8e-278c-4924-a270-089b27b4ad82 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.980391] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 902.980391] env[63372]: value = "task-1024392" [ 902.980391] env[63372]: _type = "Task" [ 902.980391] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.984126] env[63372]: DEBUG nova.network.neutron [req-1f7b2269-3bd1-40cc-8a9c-fcb8f33f714b req-91542834-d231-4e16-80d7-c6e7d3eaf739 service nova] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Updated VIF entry in instance network info cache for port 2454845a-285a-4d06-8f95-123e004fae02. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 902.984480] env[63372]: DEBUG nova.network.neutron [req-1f7b2269-3bd1-40cc-8a9c-fcb8f33f714b req-91542834-d231-4e16-80d7-c6e7d3eaf739 service nova] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Updating instance_info_cache with network_info: [{"id": "2454845a-285a-4d06-8f95-123e004fae02", "address": "fa:16:3e:83:4b:f8", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2454845a-28", "ovs_interfaceid": "2454845a-285a-4d06-8f95-123e004fae02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.990554] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024392, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.014031] env[63372]: DEBUG oslo_vmware.api [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Task: {'id': task-1024389, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.231478} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.014031] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.014259] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 903.014466] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 903.014682] env[63372]: INFO nova.compute.manager [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Took 1.17 seconds to destroy the instance on the hypervisor. [ 903.015021] env[63372]: DEBUG oslo.service.loopingcall [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.015271] env[63372]: DEBUG nova.compute.manager [-] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 903.015397] env[63372]: DEBUG nova.network.neutron [-] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 903.051399] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024390, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.371630] env[63372]: DEBUG nova.network.neutron [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Successfully created port: a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 903.403550] env[63372]: DEBUG nova.network.neutron [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance_info_cache with network_info: [{"id": "e1486286-9147-4e5e-88c1-60893a3338dd", "address": "fa:16:3e:6c:92:ac", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1486286-91", "ovs_interfaceid": "e1486286-9147-4e5e-88c1-60893a3338dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.436163] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.862s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.438591] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.170s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.438786] env[63372]: DEBUG nova.objects.instance [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63372) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 903.448580] env[63372]: DEBUG nova.network.neutron [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 903.467283] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024391, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064881} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.467283] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 903.468121] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9245e28-37ac-492b-94a9-d28b1eef85e6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.497997] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] d568b727-7836-4b1d-9f27-f159227e46f7/d568b727-7836-4b1d-9f27-f159227e46f7.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.499271] env[63372]: INFO nova.scheduler.client.report [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Deleted allocations for instance ac90a156-be00-4f62-a76e-e08914531167 [ 903.506451] env[63372]: DEBUG oslo_concurrency.lockutils [req-1f7b2269-3bd1-40cc-8a9c-fcb8f33f714b req-91542834-d231-4e16-80d7-c6e7d3eaf739 service nova] Releasing lock "refresh_cache-d568b727-7836-4b1d-9f27-f159227e46f7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.507657] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5a691d4-d19e-47cb-9c39-ea39d4247630 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.533567] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024392, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.535304] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 903.535304] env[63372]: value = "task-1024393" [ 903.535304] env[63372]: _type = "Task" [ 903.535304] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.548271] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024393, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.557740] env[63372]: DEBUG oslo_vmware.api [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024390, 'name': PowerOnVM_Task, 'duration_secs': 0.648033} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.557740] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 903.557740] env[63372]: INFO nova.compute.manager [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Took 8.84 seconds to spawn the instance on the hypervisor. [ 903.557740] env[63372]: DEBUG nova.compute.manager [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.557740] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca543af6-fa27-4ce3-8558-88c6073f4953 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.871407] env[63372]: DEBUG nova.network.neutron [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Updating instance_info_cache with network_info: [{"id": "8321625e-abd2-4cc3-b545-2190d9759158", "address": "fa:16:3e:d8:f4:21", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8321625e-ab", "ovs_interfaceid": "8321625e-abd2-4cc3-b545-2190d9759158", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.907044] env[63372]: DEBUG oslo_concurrency.lockutils [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.907044] env[63372]: DEBUG nova.objects.instance [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'migration_context' on Instance uuid 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.995025] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024392, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.667952} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.995395] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] ba1daa0a-dcf8-4586-9964-a962fcc94196/ba1daa0a-dcf8-4586-9964-a962fcc94196.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 903.995589] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.995854] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-51a6f77f-4c84-4eb7-af12-e9138cc83bf1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.004225] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 904.004225] env[63372]: value = "task-1024394" [ 904.004225] env[63372]: _type = "Task" [ 904.004225] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.012227] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024394, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.030878] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c0219f0a-cfe5-4943-b71d-b326f1b58ef9 tempest-VolumesAdminNegativeTest-2007005101 tempest-VolumesAdminNegativeTest-2007005101-project-member] Lock "ac90a156-be00-4f62-a76e-e08914531167" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.649s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.044892] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024393, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.074195] env[63372]: INFO nova.compute.manager [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Took 37.07 seconds to build instance. [ 904.096512] env[63372]: DEBUG nova.network.neutron [-] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.374487] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "refresh_cache-326a9155-1c00-4c18-a9a7-6124e709c20b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.374836] env[63372]: DEBUG nova.compute.manager [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Instance network_info: |[{"id": "8321625e-abd2-4cc3-b545-2190d9759158", "address": "fa:16:3e:d8:f4:21", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8321625e-ab", "ovs_interfaceid": "8321625e-abd2-4cc3-b545-2190d9759158", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 904.375704] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:f4:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d4ef133-b6f3-41d1-add4-92a1482195cf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8321625e-abd2-4cc3-b545-2190d9759158', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 904.383613] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Creating folder: Project (dac773c2536745e285181426ae34bb96). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 904.383912] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c019c6d7-f03e-4307-b311-b1e1e972b669 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.395228] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Created folder: Project (dac773c2536745e285181426ae34bb96) in parent group-v227230. [ 904.395228] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Creating folder: Instances. Parent ref: group-v227403. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 904.395228] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a31e6629-4ac7-41cf-b6d7-f93dc814ad6a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.403440] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Created folder: Instances in parent group-v227403. [ 904.403500] env[63372]: DEBUG oslo.service.loopingcall [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.403644] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 904.403845] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d74d8dea-bd2a-41e9-a3cb-8e1a59f4dacc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.418630] env[63372]: DEBUG nova.objects.base [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Object Instance<19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5> lazy-loaded attributes: info_cache,migration_context {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 904.419510] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22cf4724-2daf-40e8-b90f-46fe9d3319d9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.439806] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5fd66ea-be7f-4eeb-8ea7-5bbd0f2c4692 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.442106] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 904.442106] env[63372]: value = "task-1024397" [ 904.442106] env[63372]: _type = "Task" [ 904.442106] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.446435] env[63372]: DEBUG oslo_vmware.api [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 904.446435] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f32529-b850-6e5c-0213-8cea4203fbdf" [ 904.446435] env[63372]: _type = "Task" [ 904.446435] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.447835] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d268901-7e2d-4378-9b1c-9b7fb1b841f9 tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.455600] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.257s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.455600] env[63372]: DEBUG nova.objects.instance [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Lazy-loading 'resources' on Instance uuid da8cf1ff-d977-4fd1-9239-043f07bed875 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.456184] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024397, 'name': CreateVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.461091] env[63372]: DEBUG oslo_vmware.api [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f32529-b850-6e5c-0213-8cea4203fbdf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.496509] env[63372]: DEBUG nova.compute.manager [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Received event network-vif-plugged-8321625e-abd2-4cc3-b545-2190d9759158 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.496740] env[63372]: DEBUG oslo_concurrency.lockutils [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] Acquiring lock "326a9155-1c00-4c18-a9a7-6124e709c20b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.496978] env[63372]: DEBUG oslo_concurrency.lockutils [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] Lock "326a9155-1c00-4c18-a9a7-6124e709c20b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.497184] env[63372]: DEBUG oslo_concurrency.lockutils [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] Lock "326a9155-1c00-4c18-a9a7-6124e709c20b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.497364] env[63372]: DEBUG nova.compute.manager [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] No waiting events found dispatching network-vif-plugged-8321625e-abd2-4cc3-b545-2190d9759158 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 904.497531] env[63372]: WARNING nova.compute.manager [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Received unexpected event network-vif-plugged-8321625e-abd2-4cc3-b545-2190d9759158 for instance with vm_state building and task_state spawning. [ 904.497698] env[63372]: DEBUG nova.compute.manager [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Received event network-changed-8321625e-abd2-4cc3-b545-2190d9759158 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.497854] env[63372]: DEBUG nova.compute.manager [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Refreshing instance network info cache due to event network-changed-8321625e-abd2-4cc3-b545-2190d9759158. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 904.498057] env[63372]: DEBUG oslo_concurrency.lockutils [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] Acquiring lock "refresh_cache-326a9155-1c00-4c18-a9a7-6124e709c20b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.498201] env[63372]: DEBUG oslo_concurrency.lockutils [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] Acquired lock "refresh_cache-326a9155-1c00-4c18-a9a7-6124e709c20b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.498364] env[63372]: DEBUG nova.network.neutron [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Refreshing network info cache for port 8321625e-abd2-4cc3-b545-2190d9759158 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 904.516191] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024394, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075389} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.516605] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.517725] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c06eeb-f37f-4978-82a0-64a8c62cf05e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.540046] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] ba1daa0a-dcf8-4586-9964-a962fcc94196/ba1daa0a-dcf8-4586-9964-a962fcc94196.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.540829] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e2ba2ee-5629-4681-b3d0-cdc542638e8b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.562922] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024393, 'name': ReconfigVM_Task, 'duration_secs': 0.600599} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.564183] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Reconfigured VM instance instance-00000051 to attach disk [datastore2] d568b727-7836-4b1d-9f27-f159227e46f7/d568b727-7836-4b1d-9f27-f159227e46f7.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.564860] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 904.564860] env[63372]: value = "task-1024398" [ 904.564860] env[63372]: _type = "Task" [ 904.564860] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.565070] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c6924f6-910d-44ba-834e-10e939341c6f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.574810] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024398, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.576150] env[63372]: DEBUG oslo_concurrency.lockutils [None req-eea9591c-f4f5-45e6-ae81-09332689ce76 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "75560852-8c6d-4220-b251-81e8721a4b7b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.585s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.576441] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 904.576441] env[63372]: value = "task-1024399" [ 904.576441] env[63372]: _type = "Task" [ 904.576441] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.583963] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024399, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.599162] env[63372]: INFO nova.compute.manager [-] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Took 1.58 seconds to deallocate network for instance. [ 904.954488] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024397, 'name': CreateVM_Task, 'duration_secs': 0.342908} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.955013] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 904.955982] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.956198] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.956573] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.962334] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88ec46b8-6324-4936-ac8f-4d58d67373f9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.964102] env[63372]: DEBUG oslo_vmware.api [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f32529-b850-6e5c-0213-8cea4203fbdf, 'name': SearchDatastore_Task, 'duration_secs': 0.014669} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.964645] env[63372]: DEBUG oslo_concurrency.lockutils [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.967587] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 904.967587] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52302f0b-d3bd-650d-8811-0fd91a7e4ff6" [ 904.967587] env[63372]: _type = "Task" [ 904.967587] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.975418] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52302f0b-d3bd-650d-8811-0fd91a7e4ff6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.078478] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024398, 'name': ReconfigVM_Task, 'duration_secs': 0.486885} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.083265] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Reconfigured VM instance instance-00000050 to attach disk [datastore1] ba1daa0a-dcf8-4586-9964-a962fcc94196/ba1daa0a-dcf8-4586-9964-a962fcc94196.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.083904] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0811c1e5-7c09-4e03-8d9f-d701710c3ff8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.094727] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024399, 'name': Rename_Task, 'duration_secs': 0.186941} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.095983] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 905.096324] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 905.096324] env[63372]: value = "task-1024400" [ 905.096324] env[63372]: _type = "Task" [ 905.096324] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.096513] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-11362f2c-7849-4ec4-b0d0-47babeeab686 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.106167] env[63372]: DEBUG oslo_concurrency.lockutils [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.111067] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024400, 'name': Rename_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.112339] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 905.112339] env[63372]: value = "task-1024401" [ 905.112339] env[63372]: _type = "Task" [ 905.112339] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.119714] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024401, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.190931] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3600fb7d-cfd3-417b-95fb-964a773f1599 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.200860] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf29141-46fd-4103-9463-bae93cf1c43d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.240732] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc16bfdb-cc7f-497b-b796-c4d8e3056927 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.243718] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "75560852-8c6d-4220-b251-81e8721a4b7b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.244040] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "75560852-8c6d-4220-b251-81e8721a4b7b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.244291] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "75560852-8c6d-4220-b251-81e8721a4b7b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.244507] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "75560852-8c6d-4220-b251-81e8721a4b7b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.244707] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "75560852-8c6d-4220-b251-81e8721a4b7b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.247463] env[63372]: INFO nova.compute.manager [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Terminating instance [ 905.249762] env[63372]: DEBUG nova.compute.manager [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 905.250056] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 905.250851] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd47844f-f946-4d68-aff4-8c6e16c3b13f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.258424] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecde650c-eaa8-4459-8ce3-68b771a1f828 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.263909] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 905.264514] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b74385cf-ce6a-46f3-bcce-13b19ebbeb41 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.274403] env[63372]: DEBUG nova.compute.provider_tree [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.276804] env[63372]: DEBUG oslo_vmware.api [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 905.276804] env[63372]: value = "task-1024402" [ 905.276804] env[63372]: _type = "Task" [ 905.276804] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.284910] env[63372]: DEBUG oslo_vmware.api [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024402, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.380997] env[63372]: DEBUG nova.network.neutron [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Successfully updated port: a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 905.435788] env[63372]: DEBUG nova.network.neutron [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Updated VIF entry in instance network info cache for port 8321625e-abd2-4cc3-b545-2190d9759158. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 905.435788] env[63372]: DEBUG nova.network.neutron [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Updating instance_info_cache with network_info: [{"id": "8321625e-abd2-4cc3-b545-2190d9759158", "address": "fa:16:3e:d8:f4:21", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8321625e-ab", "ovs_interfaceid": "8321625e-abd2-4cc3-b545-2190d9759158", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.478355] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52302f0b-d3bd-650d-8811-0fd91a7e4ff6, 'name': SearchDatastore_Task, 'duration_secs': 0.008697} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.478809] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.479126] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.479371] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.479524] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.479707] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.479979] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6508f82c-cfbf-42d7-99d0-0205d2e5ac35 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.488232] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.488346] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 905.489114] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6f5290e-36d9-469e-ab6f-921083c3df64 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.494717] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 905.494717] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f9eff1-d4f1-1fa7-c71a-9c7987ed66a2" [ 905.494717] env[63372]: _type = "Task" [ 905.494717] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.501739] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f9eff1-d4f1-1fa7-c71a-9c7987ed66a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.608026] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024400, 'name': Rename_Task, 'duration_secs': 0.213238} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.608026] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 905.608202] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bdfd8f28-84d4-49ed-908a-4197a398b497 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.616710] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 905.616710] env[63372]: value = "task-1024403" [ 905.616710] env[63372]: _type = "Task" [ 905.616710] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.622665] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024401, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.627024] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024403, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.779788] env[63372]: DEBUG nova.scheduler.client.report [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.792597] env[63372]: DEBUG oslo_concurrency.lockutils [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquiring lock "662d9fdf-ec0d-4466-8d15-abc942e4dad8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.792804] env[63372]: DEBUG oslo_concurrency.lockutils [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Lock "662d9fdf-ec0d-4466-8d15-abc942e4dad8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.797393] env[63372]: DEBUG oslo_vmware.api [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024402, 'name': PowerOffVM_Task, 'duration_secs': 0.153359} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.797901] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 905.798530] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 905.798530] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fdf4b0da-9a72-4b8b-a6d7-1bc4cd82236e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.863457] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 905.863786] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 905.864058] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleting the datastore file [datastore1] 75560852-8c6d-4220-b251-81e8721a4b7b {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 905.864414] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7be95043-d8da-4cdd-ae6c-1f4d70de305b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.874994] env[63372]: DEBUG oslo_vmware.api [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 905.874994] env[63372]: value = "task-1024405" [ 905.874994] env[63372]: _type = "Task" [ 905.874994] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.885971] env[63372]: DEBUG oslo_vmware.api [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024405, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.887701] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.887871] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.888057] env[63372]: DEBUG nova.network.neutron [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 905.926660] env[63372]: DEBUG nova.compute.manager [req-05ae4d5b-2e26-435d-a38b-028a65ba1688 req-d5a5c555-0251-48c2-97c9-c4dda551e617 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Received event network-vif-plugged-a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.926859] env[63372]: DEBUG oslo_concurrency.lockutils [req-05ae4d5b-2e26-435d-a38b-028a65ba1688 req-d5a5c555-0251-48c2-97c9-c4dda551e617 service nova] Acquiring lock "13bde897-8446-42a2-b02d-2f5b48e6f432-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.927094] env[63372]: DEBUG oslo_concurrency.lockutils [req-05ae4d5b-2e26-435d-a38b-028a65ba1688 req-d5a5c555-0251-48c2-97c9-c4dda551e617 service nova] Lock "13bde897-8446-42a2-b02d-2f5b48e6f432-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.927274] env[63372]: DEBUG oslo_concurrency.lockutils [req-05ae4d5b-2e26-435d-a38b-028a65ba1688 req-d5a5c555-0251-48c2-97c9-c4dda551e617 service nova] Lock "13bde897-8446-42a2-b02d-2f5b48e6f432-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.927724] env[63372]: DEBUG nova.compute.manager [req-05ae4d5b-2e26-435d-a38b-028a65ba1688 req-d5a5c555-0251-48c2-97c9-c4dda551e617 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] No waiting events found dispatching network-vif-plugged-a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 905.927724] env[63372]: WARNING nova.compute.manager [req-05ae4d5b-2e26-435d-a38b-028a65ba1688 req-d5a5c555-0251-48c2-97c9-c4dda551e617 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Received unexpected event network-vif-plugged-a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c for instance with vm_state active and task_state None. [ 905.938349] env[63372]: DEBUG oslo_concurrency.lockutils [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] Releasing lock "refresh_cache-326a9155-1c00-4c18-a9a7-6124e709c20b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.938703] env[63372]: DEBUG nova.compute.manager [req-8146421c-ad54-402f-b1cd-0c10b9d650b1 req-2c25e24d-4f28-4125-af16-76fca23c6f5f service nova] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Received event network-vif-deleted-777f211d-71d8-4f32-bfac-588b17d506f9 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.005778] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f9eff1-d4f1-1fa7-c71a-9c7987ed66a2, 'name': SearchDatastore_Task, 'duration_secs': 0.008515} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.006689] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7d3623a-6674-4e77-8e3c-1e8ea21affb3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.012605] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 906.012605] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5261f7f1-8e1d-87aa-5fe7-0375d6eb33d1" [ 906.012605] env[63372]: _type = "Task" [ 906.012605] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.023191] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5261f7f1-8e1d-87aa-5fe7-0375d6eb33d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.126365] env[63372]: DEBUG oslo_vmware.api [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024401, 'name': PowerOnVM_Task, 'duration_secs': 0.82601} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.127448] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 906.127684] env[63372]: INFO nova.compute.manager [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Took 6.73 seconds to spawn the instance on the hypervisor. [ 906.127873] env[63372]: DEBUG nova.compute.manager [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.128963] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51eaadcf-2e49-4801-9600-c44f5b77b842 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.134352] env[63372]: DEBUG oslo_vmware.api [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024403, 'name': PowerOnVM_Task, 'duration_secs': 0.471617} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.135051] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 906.135284] env[63372]: INFO nova.compute.manager [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Took 9.08 seconds to spawn the instance on the hypervisor. [ 906.135463] env[63372]: DEBUG nova.compute.manager [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.136344] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5aad2cf-ea1e-470b-9bb6-34d36d9aa53e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.290464] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.835s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.292852] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.005s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.293113] env[63372]: DEBUG nova.objects.instance [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Lazy-loading 'resources' on Instance uuid 1f52f9ee-2af3-4321-8ec6-5d5e841ed880 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.298917] env[63372]: DEBUG nova.compute.manager [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 906.385524] env[63372]: DEBUG oslo_vmware.api [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024405, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.214882} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.385870] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 906.386091] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 906.386320] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 906.386526] env[63372]: INFO nova.compute.manager [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 906.386773] env[63372]: DEBUG oslo.service.loopingcall [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.386996] env[63372]: DEBUG nova.compute.manager [-] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 906.387147] env[63372]: DEBUG nova.network.neutron [-] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 906.462595] env[63372]: WARNING nova.network.neutron [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] e1f7ca1b-541e-468f-a439-a6841fc781ba already exists in list: networks containing: ['e1f7ca1b-541e-468f-a439-a6841fc781ba']. ignoring it [ 906.524559] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5261f7f1-8e1d-87aa-5fe7-0375d6eb33d1, 'name': SearchDatastore_Task, 'duration_secs': 0.018574} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.525072] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.525170] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 326a9155-1c00-4c18-a9a7-6124e709c20b/326a9155-1c00-4c18-a9a7-6124e709c20b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 906.525532] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-145ec9b7-d235-4ffa-9259-39648c2a1461 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.532638] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 906.532638] env[63372]: value = "task-1024406" [ 906.532638] env[63372]: _type = "Task" [ 906.532638] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.543101] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024406, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.658916] env[63372]: INFO nova.compute.manager [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Took 39.46 seconds to build instance. [ 906.669737] env[63372]: INFO nova.compute.manager [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Took 38.45 seconds to build instance. [ 906.676589] env[63372]: INFO nova.scheduler.client.report [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Deleted allocations for instance da8cf1ff-d977-4fd1-9239-043f07bed875 [ 906.830274] env[63372]: DEBUG oslo_concurrency.lockutils [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.045034] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024406, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.047210] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f185fdc-b2bb-4d54-b1fb-94945c883e04 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.057555] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a507452b-6238-4fee-b665-086bd57a7991 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.092627] env[63372]: DEBUG nova.network.neutron [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Updating instance_info_cache with network_info: [{"id": "b02b2fb0-9262-4f24-a677-c4f61362410a", "address": "fa:16:3e:03:c1:fa", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb02b2fb0-92", "ovs_interfaceid": "b02b2fb0-9262-4f24-a677-c4f61362410a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c", "address": "fa:16:3e:82:b5:73", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5e3b219-68", "ovs_interfaceid": "a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.092627] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52390a7d-d966-4ef8-bb5d-d9b2b4243d1c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.105517] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e18bb64-b6aa-43ee-aff9-a4cba7402880 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.119340] env[63372]: DEBUG nova.compute.provider_tree [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.171053] env[63372]: DEBUG oslo_concurrency.lockutils [None req-4802e32f-5f62-499e-b678-c498f3bfcec8 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "ba1daa0a-dcf8-4586-9964-a962fcc94196" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.980s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.171423] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1679965c-4f12-42fe-8bce-897019cf6447 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "d568b727-7836-4b1d-9f27-f159227e46f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.968s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.187381] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3e99a26a-23f2-4143-aca9-c2809674c76b tempest-ServerShowV257Test-1535620152 tempest-ServerShowV257Test-1535620152-project-member] Lock "da8cf1ff-d977-4fd1-9239-043f07bed875" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.676s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.469246] env[63372]: DEBUG nova.network.neutron [-] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.544302] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024406, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.687976} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.544619] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 326a9155-1c00-4c18-a9a7-6124e709c20b/326a9155-1c00-4c18-a9a7-6124e709c20b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 907.544886] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 907.545188] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-465a8f46-5199-4b76-bfb0-243e016d5da9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.551836] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 907.551836] env[63372]: value = "task-1024407" [ 907.551836] env[63372]: _type = "Task" [ 907.551836] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.560870] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024407, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.597304] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.597966] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.598142] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.598973] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb69191a-27ae-4d2b-8f60-02c40a1694d0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.615139] env[63372]: DEBUG nova.virt.hardware [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 907.615384] env[63372]: DEBUG nova.virt.hardware [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 907.615536] env[63372]: DEBUG nova.virt.hardware [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 907.615720] env[63372]: DEBUG nova.virt.hardware [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 907.615852] env[63372]: DEBUG nova.virt.hardware [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 907.616021] env[63372]: DEBUG nova.virt.hardware [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 907.616229] env[63372]: DEBUG nova.virt.hardware [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 907.616389] env[63372]: DEBUG nova.virt.hardware [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 907.616555] env[63372]: DEBUG nova.virt.hardware [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 907.616721] env[63372]: DEBUG nova.virt.hardware [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 907.616897] env[63372]: DEBUG nova.virt.hardware [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 907.623158] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Reconfiguring VM to attach interface {{(pid=63372) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 907.624424] env[63372]: DEBUG nova.scheduler.client.report [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.627410] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6bcfa602-e9dd-4df3-a61e-bbbb5df134e4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.647022] env[63372]: DEBUG oslo_vmware.api [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 907.647022] env[63372]: value = "task-1024408" [ 907.647022] env[63372]: _type = "Task" [ 907.647022] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.654081] env[63372]: DEBUG oslo_vmware.api [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024408, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.975294] env[63372]: INFO nova.compute.manager [-] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Took 1.59 seconds to deallocate network for instance. [ 908.015522] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "95178e2e-9d71-4606-a856-c530bfbb9345" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.016033] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "95178e2e-9d71-4606-a856-c530bfbb9345" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.062574] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024407, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.384834} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.063644] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 908.063644] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddca605d-1501-4f73-a8d9-84cf0a07b6f2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.086199] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 326a9155-1c00-4c18-a9a7-6124e709c20b/326a9155-1c00-4c18-a9a7-6124e709c20b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.087148] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e65cf08-439a-4367-b4bb-ce26a1b7a577 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.107176] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 908.107176] env[63372]: value = "task-1024409" [ 908.107176] env[63372]: _type = "Task" [ 908.107176] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.115669] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024409, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.140849] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.848s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.143281] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.464s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.143359] env[63372]: DEBUG nova.objects.instance [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lazy-loading 'resources' on Instance uuid c0b9bd7e-92e2-44a1-b461-e54c4411ddc8 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.155128] env[63372]: DEBUG oslo_vmware.api [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024408, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.188437] env[63372]: INFO nova.scheduler.client.report [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Deleted allocations for instance 1f52f9ee-2af3-4321-8ec6-5d5e841ed880 [ 908.487402] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.519139] env[63372]: DEBUG nova.compute.manager [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 908.621469] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.659418] env[63372]: DEBUG oslo_vmware.api [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024408, 'name': ReconfigVM_Task, 'duration_secs': 0.987957} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.661021] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.661021] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Reconfigured VM to attach interface {{(pid=63372) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 908.681028] env[63372]: DEBUG nova.compute.manager [req-98e8f7cc-4cbe-4a8e-925a-26a624ade617 req-f2457ae2-348f-4471-99bb-72d7af10e49f service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Received event network-changed-a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.681189] env[63372]: DEBUG nova.compute.manager [req-98e8f7cc-4cbe-4a8e-925a-26a624ade617 req-f2457ae2-348f-4471-99bb-72d7af10e49f service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Refreshing instance network info cache due to event network-changed-a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 908.681398] env[63372]: DEBUG oslo_concurrency.lockutils [req-98e8f7cc-4cbe-4a8e-925a-26a624ade617 req-f2457ae2-348f-4471-99bb-72d7af10e49f service nova] Acquiring lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.681548] env[63372]: DEBUG oslo_concurrency.lockutils [req-98e8f7cc-4cbe-4a8e-925a-26a624ade617 req-f2457ae2-348f-4471-99bb-72d7af10e49f service nova] Acquired lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.681713] env[63372]: DEBUG nova.network.neutron [req-98e8f7cc-4cbe-4a8e-925a-26a624ade617 req-f2457ae2-348f-4471-99bb-72d7af10e49f service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Refreshing network info cache for port a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 908.703358] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f166e257-a0ff-40d3-a9f3-fe42e22827c6 tempest-ServerRescueTestJSONUnderV235-966483274 tempest-ServerRescueTestJSONUnderV235-966483274-project-member] Lock "1f52f9ee-2af3-4321-8ec6-5d5e841ed880" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.371s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.908218] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d87f1b7-25a6-43e9-ad98-bf52aa342d38 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.916610] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35a59db-5ad8-4225-a547-2775cfda8d4e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.949755] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c630707-fe5d-4ddf-ac03-d374916803ea {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.956290] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a75b0d-c58a-46c6-9206-915602820e00 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.971922] env[63372]: DEBUG nova.compute.provider_tree [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.046935] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.120198] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024409, 'name': ReconfigVM_Task, 'duration_secs': 0.809664} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.120515] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 326a9155-1c00-4c18-a9a7-6124e709c20b/326a9155-1c00-4c18-a9a7-6124e709c20b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.121170] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3844d1c-a2ac-4d2b-afd3-d78a3ef0b433 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.129030] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 909.129030] env[63372]: value = "task-1024410" [ 909.129030] env[63372]: _type = "Task" [ 909.129030] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.137800] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024410, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.166953] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ae6de7f4-032b-43c1-85d1-07f7ccd106b1 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-13bde897-8446-42a2-b02d-2f5b48e6f432-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.496s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.476327] env[63372]: DEBUG nova.scheduler.client.report [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.504204] env[63372]: DEBUG nova.network.neutron [req-98e8f7cc-4cbe-4a8e-925a-26a624ade617 req-f2457ae2-348f-4471-99bb-72d7af10e49f service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Updated VIF entry in instance network info cache for port a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 909.504681] env[63372]: DEBUG nova.network.neutron [req-98e8f7cc-4cbe-4a8e-925a-26a624ade617 req-f2457ae2-348f-4471-99bb-72d7af10e49f service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Updating instance_info_cache with network_info: [{"id": "b02b2fb0-9262-4f24-a677-c4f61362410a", "address": "fa:16:3e:03:c1:fa", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb02b2fb0-92", "ovs_interfaceid": "b02b2fb0-9262-4f24-a677-c4f61362410a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c", "address": "fa:16:3e:82:b5:73", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5e3b219-68", "ovs_interfaceid": "a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.605377] env[63372]: INFO nova.compute.manager [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Rebuilding instance [ 909.642374] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024410, 'name': Rename_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.663655] env[63372]: DEBUG nova.compute.manager [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.664858] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd02513-d4e2-4ef7-8a82-f241c765e07f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.981091] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.838s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.983481] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.160s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.983709] env[63372]: DEBUG nova.objects.instance [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lazy-loading 'pci_requests' on Instance uuid c01a5d24-eb46-4a69-993e-753880ce8e85 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.008296] env[63372]: DEBUG oslo_concurrency.lockutils [req-98e8f7cc-4cbe-4a8e-925a-26a624ade617 req-f2457ae2-348f-4471-99bb-72d7af10e49f service nova] Releasing lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.008754] env[63372]: DEBUG nova.compute.manager [req-98e8f7cc-4cbe-4a8e-925a-26a624ade617 req-f2457ae2-348f-4471-99bb-72d7af10e49f service nova] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Received event network-vif-deleted-43013e6e-28ee-4f8c-afa8-20c12b01c0f8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.012950] env[63372]: INFO nova.scheduler.client.report [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Deleted allocations for instance c0b9bd7e-92e2-44a1-b461-e54c4411ddc8 [ 910.142133] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024410, 'name': Rename_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.177058] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 910.177420] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c001baa4-693c-4493-bb5c-ecd7b372d699 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.184569] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 910.184569] env[63372]: value = "task-1024411" [ 910.184569] env[63372]: _type = "Task" [ 910.184569] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.195210] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024411, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.492354] env[63372]: DEBUG nova.objects.instance [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lazy-loading 'numa_topology' on Instance uuid c01a5d24-eb46-4a69-993e-753880ce8e85 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.526796] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d8664604-8050-434d-9c72-31acb20c02ef tempest-ImagesTestJSON-1251160395 tempest-ImagesTestJSON-1251160395-project-member] Lock "c0b9bd7e-92e2-44a1-b461-e54c4411ddc8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.725s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.644939] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024410, 'name': Rename_Task, 'duration_secs': 1.226571} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.644939] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 910.644939] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9b7c2aa-8acd-4544-bef7-23f5e6adc6c8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.652028] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 910.652028] env[63372]: value = "task-1024412" [ 910.652028] env[63372]: _type = "Task" [ 910.652028] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.659504] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024412, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.695936] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024411, 'name': PowerOffVM_Task, 'duration_secs': 0.496542} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.699023] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 910.699023] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 910.699023] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6191c0e-08a3-4733-bcf0-4e2213bd2d1a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.705509] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 910.705901] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e6cd0228-8f32-4cf6-84ac-c036c7035f80 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.887021] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 910.887021] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 910.887021] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleting the datastore file [datastore2] d568b727-7836-4b1d-9f27-f159227e46f7 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.887021] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4546050e-c852-4d37-a41b-ee5bc0c98f74 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.892857] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 910.892857] env[63372]: value = "task-1024414" [ 910.892857] env[63372]: _type = "Task" [ 910.892857] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.902997] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024414, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.006652] env[63372]: INFO nova.compute.claims [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 911.043651] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "interface-13bde897-8446-42a2-b02d-2f5b48e6f432-eac444f7-19cf-4cc8-b354-21b08d90753c" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.043959] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-13bde897-8446-42a2-b02d-2f5b48e6f432-eac444f7-19cf-4cc8-b354-21b08d90753c" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.044345] env[63372]: DEBUG nova.objects.instance [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'flavor' on Instance uuid 13bde897-8446-42a2-b02d-2f5b48e6f432 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.160235] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024412, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.405837] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024414, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.665868] env[63372]: DEBUG oslo_vmware.api [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024412, 'name': PowerOnVM_Task, 'duration_secs': 1.01083} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.666704] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 911.666983] env[63372]: INFO nova.compute.manager [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Took 9.91 seconds to spawn the instance on the hypervisor. [ 911.667191] env[63372]: DEBUG nova.compute.manager [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.667978] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06bb6f54-2aa3-4ad7-9aa1-ffac6513554d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.816774] env[63372]: DEBUG nova.objects.instance [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'pci_requests' on Instance uuid 13bde897-8446-42a2-b02d-2f5b48e6f432 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.903599] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024414, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.595744} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.903857] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.904053] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 911.904237] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 912.196275] env[63372]: INFO nova.compute.manager [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Took 43.65 seconds to build instance. [ 912.287555] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2398243c-bf3f-4ea5-b485-2a7afdc42699 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.296445] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6aef30-da0d-4d6f-b04d-29e88e2a2f15 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.333972] env[63372]: DEBUG nova.objects.base [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Object Instance<13bde897-8446-42a2-b02d-2f5b48e6f432> lazy-loaded attributes: flavor,pci_requests {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 912.334263] env[63372]: DEBUG nova.network.neutron [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 912.337248] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9368d5d4-a4a7-4091-896a-65b1c48c734b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.345076] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8527461-6d3b-492f-aec8-6edf44fd598d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.359817] env[63372]: DEBUG nova.compute.provider_tree [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.476592] env[63372]: DEBUG nova.policy [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7ffe6776b664c00b9086046f9171ec6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46d2cb188a004dc385f2aa15458be929', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 912.700515] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30151f14-75d0-4ce7-a245-3763070150ff tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "326a9155-1c00-4c18-a9a7-6124e709c20b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.160s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.864273] env[63372]: DEBUG nova.scheduler.client.report [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.947583] env[63372]: DEBUG nova.virt.hardware [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.947583] env[63372]: DEBUG nova.virt.hardware [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.947583] env[63372]: DEBUG nova.virt.hardware [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.947583] env[63372]: DEBUG nova.virt.hardware [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.947825] env[63372]: DEBUG nova.virt.hardware [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.948591] env[63372]: DEBUG nova.virt.hardware [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.948591] env[63372]: DEBUG nova.virt.hardware [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.948591] env[63372]: DEBUG nova.virt.hardware [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.948591] env[63372]: DEBUG nova.virt.hardware [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.948820] env[63372]: DEBUG nova.virt.hardware [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.949239] env[63372]: DEBUG nova.virt.hardware [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.949962] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36552aa1-11f3-4cfc-bcd0-bbf5837a17df {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.958700] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610b7725-e769-4d7d-9b77-7cb382129110 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.974120] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:4b:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d5d0e0d-cdec-474a-a891-a9ceff15a8b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2454845a-285a-4d06-8f95-123e004fae02', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.982724] env[63372]: DEBUG oslo.service.loopingcall [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.983121] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 912.983324] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9ed12268-37de-4b73-9558-ef5d087ed623 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.001997] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 913.001997] env[63372]: value = "task-1024415" [ 913.001997] env[63372]: _type = "Task" [ 913.001997] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.013703] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024415, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.369555] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.386s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.371663] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.476s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.371663] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.371824] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63372) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 913.372202] env[63372]: DEBUG oslo_concurrency.lockutils [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 8.408s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.376880] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac2f99be-2ad3-4822-b001-c4c810a34052 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.383013] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1bff3bc-14c4-4c8a-9ef2-fc72f20e3165 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.402063] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab66b76f-0c90-4feb-a83a-08d9e2283067 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.406886] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b99641-a516-41bb-b77c-b4de7ad7c65f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.440583] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180965MB free_disk=185GB free_vcpus=48 pci_devices=None {{(pid=63372) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 913.440743] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.459016] env[63372]: INFO nova.network.neutron [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updating port c92e8cc9-dc72-4f20-b087-1d323c502108 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 913.513490] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024415, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.017626] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024415, 'name': CreateVM_Task, 'duration_secs': 0.622923} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.020301] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 914.021928] env[63372]: DEBUG oslo_concurrency.lockutils [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.022170] env[63372]: DEBUG oslo_concurrency.lockutils [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.022874] env[63372]: DEBUG oslo_concurrency.lockutils [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 914.023248] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72bc2fcb-91a0-4953-bf0d-662462b3d814 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.032091] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 914.032091] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52c0ead1-4a4f-1733-e15a-d2dfe80d54cc" [ 914.032091] env[63372]: _type = "Task" [ 914.032091] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.044622] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c0ead1-4a4f-1733-e15a-d2dfe80d54cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.145498] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57657ca8-1eb4-4226-a95d-b1af9372705d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.153996] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6edcd4d6-9d40-433c-830a-030b4f794fc5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.189526] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8891dfe9-bf60-471a-b093-90a75b13f87b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.198167] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2799239-cc4b-4f05-8279-03db5096f8b6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.212421] env[63372]: DEBUG nova.compute.provider_tree [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.543408] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c0ead1-4a4f-1733-e15a-d2dfe80d54cc, 'name': SearchDatastore_Task, 'duration_secs': 0.031435} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.543712] env[63372]: DEBUG oslo_concurrency.lockutils [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.544363] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.544363] env[63372]: DEBUG oslo_concurrency.lockutils [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.544363] env[63372]: DEBUG oslo_concurrency.lockutils [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.544623] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.544803] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4d8c2dd-04ec-4c96-b0da-708f1a070f71 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.553222] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.553425] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 914.554184] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b91f7989-e62e-4117-8699-a675713033c7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.559545] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 914.559545] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52c32809-e506-3806-4bfe-f701c937712e" [ 914.559545] env[63372]: _type = "Task" [ 914.559545] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.570393] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c32809-e506-3806-4bfe-f701c937712e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.715444] env[63372]: DEBUG nova.scheduler.client.report [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.807812] env[63372]: DEBUG nova.network.neutron [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Successfully updated port: eac444f7-19cf-4cc8-b354-21b08d90753c {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 914.923079] env[63372]: INFO nova.compute.manager [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Rebuilding instance [ 914.995490] env[63372]: DEBUG nova.compute.manager [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.996399] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0133730e-d1d6-4514-a9f6-77ecadc75d2a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.071267] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c32809-e506-3806-4bfe-f701c937712e, 'name': SearchDatastore_Task, 'duration_secs': 0.015813} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.072197] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50bd9f1c-a65c-49a5-b56c-481253cb6995 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.078229] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 915.078229] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5252eea0-fe31-1922-9a57-8e948bc8aec0" [ 915.078229] env[63372]: _type = "Task" [ 915.078229] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.088217] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5252eea0-fe31-1922-9a57-8e948bc8aec0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.111131] env[63372]: DEBUG nova.compute.manager [req-bd4a3f2c-dc5e-48cc-9dac-05d52fbab8e7 req-9fce1f10-5e1e-47ea-9b06-612cbfd90ec6 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Received event network-vif-plugged-eac444f7-19cf-4cc8-b354-21b08d90753c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.111368] env[63372]: DEBUG oslo_concurrency.lockutils [req-bd4a3f2c-dc5e-48cc-9dac-05d52fbab8e7 req-9fce1f10-5e1e-47ea-9b06-612cbfd90ec6 service nova] Acquiring lock "13bde897-8446-42a2-b02d-2f5b48e6f432-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.112279] env[63372]: DEBUG oslo_concurrency.lockutils [req-bd4a3f2c-dc5e-48cc-9dac-05d52fbab8e7 req-9fce1f10-5e1e-47ea-9b06-612cbfd90ec6 service nova] Lock "13bde897-8446-42a2-b02d-2f5b48e6f432-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.112279] env[63372]: DEBUG oslo_concurrency.lockutils [req-bd4a3f2c-dc5e-48cc-9dac-05d52fbab8e7 req-9fce1f10-5e1e-47ea-9b06-612cbfd90ec6 service nova] Lock "13bde897-8446-42a2-b02d-2f5b48e6f432-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.112279] env[63372]: DEBUG nova.compute.manager [req-bd4a3f2c-dc5e-48cc-9dac-05d52fbab8e7 req-9fce1f10-5e1e-47ea-9b06-612cbfd90ec6 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] No waiting events found dispatching network-vif-plugged-eac444f7-19cf-4cc8-b354-21b08d90753c {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 915.112279] env[63372]: WARNING nova.compute.manager [req-bd4a3f2c-dc5e-48cc-9dac-05d52fbab8e7 req-9fce1f10-5e1e-47ea-9b06-612cbfd90ec6 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Received unexpected event network-vif-plugged-eac444f7-19cf-4cc8-b354-21b08d90753c for instance with vm_state active and task_state None. [ 915.311293] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.311559] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.315017] env[63372]: DEBUG nova.network.neutron [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 915.432092] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.432374] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.432460] env[63372]: DEBUG nova.network.neutron [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 915.508060] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 915.508382] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a9e425f-8db2-4160-a9b3-1bc3c268031a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.515858] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 915.515858] env[63372]: value = "task-1024416" [ 915.515858] env[63372]: _type = "Task" [ 915.515858] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.529924] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024416, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.592302] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5252eea0-fe31-1922-9a57-8e948bc8aec0, 'name': SearchDatastore_Task, 'duration_secs': 0.015353} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.592741] env[63372]: DEBUG oslo_concurrency.lockutils [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.593061] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] d568b727-7836-4b1d-9f27-f159227e46f7/d568b727-7836-4b1d-9f27-f159227e46f7.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 915.593415] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b848189-68da-4cb0-b9bb-fd593a3ccf85 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.601106] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 915.601106] env[63372]: value = "task-1024417" [ 915.601106] env[63372]: _type = "Task" [ 915.601106] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.609274] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024417, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.728284] env[63372]: DEBUG oslo_concurrency.lockutils [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.356s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.728479] env[63372]: DEBUG nova.compute.manager [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=63372) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 915.733164] env[63372]: DEBUG oslo_concurrency.lockutils [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.627s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.733384] env[63372]: DEBUG nova.objects.instance [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lazy-loading 'resources' on Instance uuid 382fa76b-b930-459e-b84e-da1b9eb74b74 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.855223] env[63372]: WARNING nova.network.neutron [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] e1f7ca1b-541e-468f-a439-a6841fc781ba already exists in list: networks containing: ['e1f7ca1b-541e-468f-a439-a6841fc781ba']. ignoring it [ 915.855439] env[63372]: WARNING nova.network.neutron [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] e1f7ca1b-541e-468f-a439-a6841fc781ba already exists in list: networks containing: ['e1f7ca1b-541e-468f-a439-a6841fc781ba']. ignoring it [ 916.045410] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024416, 'name': PowerOffVM_Task, 'duration_secs': 0.400871} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.045410] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 916.045410] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 916.045410] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28118f02-25c3-4a99-b556-83c1610ddb79 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.060337] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 916.060982] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c68ba52f-5162-4502-90aa-b2635702fb94 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.112584] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024417, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.152023] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 916.152023] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 916.152023] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleting the datastore file [datastore2] 326a9155-1c00-4c18-a9a7-6124e709c20b {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.152023] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f059e85-9cbd-45ab-88cf-bdf4e58431ec {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.159172] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 916.159172] env[63372]: value = "task-1024419" [ 916.159172] env[63372]: _type = "Task" [ 916.159172] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.170420] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024419, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.303212] env[63372]: INFO nova.scheduler.client.report [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleted allocation for migration 00ae52e3-01cb-4775-870a-c0565c40be33 [ 916.490391] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a076b5c1-9e9a-4d2a-ae51-86145a46bd57 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.499192] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6dd9649-bd85-4541-937d-b9a2ede1cc34 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.535931] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18098c1d-f174-4d91-9dc1-1636b29bd836 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.541868] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "91985614-b959-401e-bb06-d67b230ee026" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.542396] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "91985614-b959-401e-bb06-d67b230ee026" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.548300] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9959eeb-500b-497f-bf3e-3b29f24a1e0d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.565336] env[63372]: DEBUG nova.compute.provider_tree [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.612384] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024417, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.871744} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.612666] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] d568b727-7836-4b1d-9f27-f159227e46f7/d568b727-7836-4b1d-9f27-f159227e46f7.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 916.612874] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 916.613186] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-720d435f-bb50-4e08-b1a1-f50ba18c2b57 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.620674] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 916.620674] env[63372]: value = "task-1024420" [ 916.620674] env[63372]: _type = "Task" [ 916.620674] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.628416] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024420, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.669044] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024419, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.780376] env[63372]: DEBUG nova.network.neutron [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updating instance_info_cache with network_info: [{"id": "c92e8cc9-dc72-4f20-b087-1d323c502108", "address": "fa:16:3e:39:9e:be", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc92e8cc9-dc", "ovs_interfaceid": "c92e8cc9-dc72-4f20-b087-1d323c502108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.810655] env[63372]: DEBUG oslo_concurrency.lockutils [None req-913c5e67-c1d2-42fd-81f3-1f989540f462 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 15.496s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.974764] env[63372]: DEBUG nova.network.neutron [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Updating instance_info_cache with network_info: [{"id": "b02b2fb0-9262-4f24-a677-c4f61362410a", "address": "fa:16:3e:03:c1:fa", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb02b2fb0-92", "ovs_interfaceid": "b02b2fb0-9262-4f24-a677-c4f61362410a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c", "address": "fa:16:3e:82:b5:73", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5e3b219-68", "ovs_interfaceid": "a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "eac444f7-19cf-4cc8-b354-21b08d90753c", "address": "fa:16:3e:1a:e3:dd", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeac444f7-19", "ovs_interfaceid": "eac444f7-19cf-4cc8-b354-21b08d90753c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.044540] env[63372]: DEBUG nova.compute.manager [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 917.069025] env[63372]: DEBUG nova.scheduler.client.report [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.137927] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024420, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.235934} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.137927] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.137927] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c22cbd-5c9f-4037-9bc9-5d0d26427f15 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.163050] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] d568b727-7836-4b1d-9f27-f159227e46f7/d568b727-7836-4b1d-9f27-f159227e46f7.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.163050] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eec979b0-f109-4a82-bf12-69010ebf784d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.191926] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024419, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.524237} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.191926] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.191926] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 917.192221] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 917.195301] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 917.195301] env[63372]: value = "task-1024421" [ 917.195301] env[63372]: _type = "Task" [ 917.195301] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.207700] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024421, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.284647] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.324609] env[63372]: DEBUG nova.virt.hardware [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='8567782ef4b625a86cfc55712b7ea075',container_format='bare',created_at=2024-09-30T11:39:58Z,direct_url=,disk_format='vmdk',id=8988d48c-9c55-4e0b-ad32-c37a4c0ae476,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1998303241-shelved',owner='f3d866ac9602470ea9f8d8159551acc8',properties=ImageMetaProps,protected=,size=31662592,status='active',tags=,updated_at=2024-09-30T11:40:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.324866] env[63372]: DEBUG nova.virt.hardware [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.325037] env[63372]: DEBUG nova.virt.hardware [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.325230] env[63372]: DEBUG nova.virt.hardware [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.325379] env[63372]: DEBUG nova.virt.hardware [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.325527] env[63372]: DEBUG nova.virt.hardware [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.325742] env[63372]: DEBUG nova.virt.hardware [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.325940] env[63372]: DEBUG nova.virt.hardware [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.327697] env[63372]: DEBUG nova.virt.hardware [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.327903] env[63372]: DEBUG nova.virt.hardware [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.328104] env[63372]: DEBUG nova.virt.hardware [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.330160] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc2ca67-59c7-4e3f-9443-76527c506be9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.340778] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1f4e1f-f47a-4742-b9f7-303b90722a46 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.357027] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:9e:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c92e8cc9-dc72-4f20-b087-1d323c502108', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.365586] env[63372]: DEBUG oslo.service.loopingcall [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.365930] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 917.366652] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c0aefba-515d-40d1-9e9f-7c1538a7c621 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.387638] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.387638] env[63372]: value = "task-1024422" [ 917.387638] env[63372]: _type = "Task" [ 917.387638] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.396126] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024422, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.481668] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.482315] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.482460] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.486538] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2280de-9aea-4d21-8677-70d6c774a45d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.510050] env[63372]: DEBUG nova.virt.hardware [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.510341] env[63372]: DEBUG nova.virt.hardware [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.510532] env[63372]: DEBUG nova.virt.hardware [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.510843] env[63372]: DEBUG nova.virt.hardware [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.510944] env[63372]: DEBUG nova.virt.hardware [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.511144] env[63372]: DEBUG nova.virt.hardware [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.511419] env[63372]: DEBUG nova.virt.hardware [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.511626] env[63372]: DEBUG nova.virt.hardware [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.511871] env[63372]: DEBUG nova.virt.hardware [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.512075] env[63372]: DEBUG nova.virt.hardware [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.512253] env[63372]: DEBUG nova.virt.hardware [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.521200] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Reconfiguring VM to attach interface {{(pid=63372) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 917.523102] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ce63881-c308-4ef5-9e57-589f0d872d4c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.539512] env[63372]: DEBUG nova.compute.manager [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Received event network-changed-eac444f7-19cf-4cc8-b354-21b08d90753c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.539512] env[63372]: DEBUG nova.compute.manager [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Refreshing instance network info cache due to event network-changed-eac444f7-19cf-4cc8-b354-21b08d90753c. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 917.539512] env[63372]: DEBUG oslo_concurrency.lockutils [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] Acquiring lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.539512] env[63372]: DEBUG oslo_concurrency.lockutils [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] Acquired lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.539512] env[63372]: DEBUG nova.network.neutron [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Refreshing network info cache for port eac444f7-19cf-4cc8-b354-21b08d90753c {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 917.549548] env[63372]: DEBUG oslo_vmware.api [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 917.549548] env[63372]: value = "task-1024423" [ 917.549548] env[63372]: _type = "Task" [ 917.549548] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.563542] env[63372]: DEBUG oslo_vmware.api [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024423, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.576084] env[63372]: DEBUG oslo_concurrency.lockutils [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.843s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.579226] env[63372]: DEBUG oslo_concurrency.lockutils [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.749s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.580745] env[63372]: INFO nova.compute.claims [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.586019] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.600199] env[63372]: INFO nova.scheduler.client.report [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Deleted allocations for instance 382fa76b-b930-459e-b84e-da1b9eb74b74 [ 917.710021] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024421, 'name': ReconfigVM_Task, 'duration_secs': 0.429244} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.710021] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Reconfigured VM instance instance-00000051 to attach disk [datastore2] d568b727-7836-4b1d-9f27-f159227e46f7/d568b727-7836-4b1d-9f27-f159227e46f7.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.710377] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0be3b01c-2bfc-4d97-8fcb-e7307a8602f7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.719174] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 917.719174] env[63372]: value = "task-1024424" [ 917.719174] env[63372]: _type = "Task" [ 917.719174] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.732116] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024424, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.896917] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024422, 'name': CreateVM_Task, 'duration_secs': 0.424957} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.897112] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 917.897850] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8988d48c-9c55-4e0b-ad32-c37a4c0ae476" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.898026] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8988d48c-9c55-4e0b-ad32-c37a4c0ae476" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.898484] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8988d48c-9c55-4e0b-ad32-c37a4c0ae476" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 917.898689] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea7be189-0e5f-4959-838d-38fb846c9cae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.903443] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 917.903443] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5268df87-a8dc-0c94-acbe-1a056b0e3030" [ 917.903443] env[63372]: _type = "Task" [ 917.903443] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.911389] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5268df87-a8dc-0c94-acbe-1a056b0e3030, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.993071] env[63372]: DEBUG nova.objects.instance [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'flavor' on Instance uuid 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.060393] env[63372]: DEBUG oslo_vmware.api [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024423, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.106720] env[63372]: DEBUG oslo_concurrency.lockutils [None req-220ae075-56e4-482a-991d-7b69343a9329 tempest-ImagesOneServerNegativeTestJSON-1936279872 tempest-ImagesOneServerNegativeTestJSON-1936279872-project-member] Lock "382fa76b-b930-459e-b84e-da1b9eb74b74" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.273s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.234689] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024424, 'name': Rename_Task, 'duration_secs': 0.233543} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.236893] env[63372]: DEBUG nova.virt.hardware [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 918.237267] env[63372]: DEBUG nova.virt.hardware [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 918.237613] env[63372]: DEBUG nova.virt.hardware [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 918.237833] env[63372]: DEBUG nova.virt.hardware [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 918.238096] env[63372]: DEBUG nova.virt.hardware [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 918.238346] env[63372]: DEBUG nova.virt.hardware [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 918.238694] env[63372]: DEBUG nova.virt.hardware [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 918.238935] env[63372]: DEBUG nova.virt.hardware [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 918.241460] env[63372]: DEBUG nova.virt.hardware [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 918.241460] env[63372]: DEBUG nova.virt.hardware [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 918.241460] env[63372]: DEBUG nova.virt.hardware [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 918.241460] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 918.242673] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de4d1e6c-a4da-4393-8942-beb4c7b9a881 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.244676] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95d0f6db-b713-4946-88c5-3f3b7e798697 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.254416] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1988141e-bd2c-473a-8d71-0c2b60dba77e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.258621] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 918.258621] env[63372]: value = "task-1024425" [ 918.258621] env[63372]: _type = "Task" [ 918.258621] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.278022] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:f4:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d4ef133-b6f3-41d1-add4-92a1482195cf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8321625e-abd2-4cc3-b545-2190d9759158', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.284570] env[63372]: DEBUG oslo.service.loopingcall [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.288242] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 918.288242] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f2b99a3-fccb-472e-9f37-0d887a1eedb7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.305198] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024425, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.314274] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.314274] env[63372]: value = "task-1024426" [ 918.314274] env[63372]: _type = "Task" [ 918.314274] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.324428] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024426, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.360334] env[63372]: DEBUG nova.network.neutron [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Updated VIF entry in instance network info cache for port eac444f7-19cf-4cc8-b354-21b08d90753c. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 918.360849] env[63372]: DEBUG nova.network.neutron [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Updating instance_info_cache with network_info: [{"id": "b02b2fb0-9262-4f24-a677-c4f61362410a", "address": "fa:16:3e:03:c1:fa", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb02b2fb0-92", "ovs_interfaceid": "b02b2fb0-9262-4f24-a677-c4f61362410a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c", "address": "fa:16:3e:82:b5:73", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5e3b219-68", "ovs_interfaceid": "a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "eac444f7-19cf-4cc8-b354-21b08d90753c", "address": "fa:16:3e:1a:e3:dd", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeac444f7-19", "ovs_interfaceid": "eac444f7-19cf-4cc8-b354-21b08d90753c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.414245] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8988d48c-9c55-4e0b-ad32-c37a4c0ae476" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.414545] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Processing image 8988d48c-9c55-4e0b-ad32-c37a4c0ae476 {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 918.414772] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8988d48c-9c55-4e0b-ad32-c37a4c0ae476/8988d48c-9c55-4e0b-ad32-c37a4c0ae476.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.414919] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8988d48c-9c55-4e0b-ad32-c37a4c0ae476/8988d48c-9c55-4e0b-ad32-c37a4c0ae476.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.415121] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 918.415378] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ab11d22-d9a4-4672-8ad0-c224887142b1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.423970] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 918.424180] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 918.424968] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa7e0c2c-946b-475d-8e76-98ff9dd77373 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.430914] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 918.430914] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52bd3895-236a-7f78-358b-261be67b2fb2" [ 918.430914] env[63372]: _type = "Task" [ 918.430914] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.439382] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52bd3895-236a-7f78-358b-261be67b2fb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.499540] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.499540] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.499799] env[63372]: DEBUG nova.network.neutron [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 918.499854] env[63372]: DEBUG nova.objects.instance [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'info_cache' on Instance uuid 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.561983] env[63372]: DEBUG oslo_vmware.api [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024423, 'name': ReconfigVM_Task, 'duration_secs': 0.64982} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.562750] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.563029] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Reconfigured VM to attach interface {{(pid=63372) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 918.755176] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7af6e8-ad9c-40a1-a17a-0cad245499b8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.765861] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e10380a1-8f9b-4202-999c-4b08bb72d368 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.776061] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024425, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.804062] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6525cede-645d-4243-8184-7dbabd89e75e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.813382] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f90ad1a-dfe1-4a8e-8de1-e9e67cd3d6b4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.832310] env[63372]: DEBUG nova.compute.provider_tree [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.836731] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024426, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.864097] env[63372]: DEBUG oslo_concurrency.lockutils [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] Releasing lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.864304] env[63372]: DEBUG nova.compute.manager [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Received event network-vif-plugged-c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.864506] env[63372]: DEBUG oslo_concurrency.lockutils [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] Acquiring lock "c01a5d24-eb46-4a69-993e-753880ce8e85-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.864712] env[63372]: DEBUG oslo_concurrency.lockutils [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.864880] env[63372]: DEBUG oslo_concurrency.lockutils [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.865056] env[63372]: DEBUG nova.compute.manager [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] No waiting events found dispatching network-vif-plugged-c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.865322] env[63372]: WARNING nova.compute.manager [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Received unexpected event network-vif-plugged-c92e8cc9-dc72-4f20-b087-1d323c502108 for instance with vm_state shelved_offloaded and task_state spawning. [ 918.865389] env[63372]: DEBUG nova.compute.manager [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Received event network-changed-c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.865538] env[63372]: DEBUG nova.compute.manager [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Refreshing instance network info cache due to event network-changed-c92e8cc9-dc72-4f20-b087-1d323c502108. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 918.865726] env[63372]: DEBUG oslo_concurrency.lockutils [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] Acquiring lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.865866] env[63372]: DEBUG oslo_concurrency.lockutils [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] Acquired lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.866034] env[63372]: DEBUG nova.network.neutron [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Refreshing network info cache for port c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 918.943937] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Preparing fetch location {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 918.944281] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Fetch image to [datastore2] OSTACK_IMG_be55a7c0-a3c7-4595-9d98-d55cc83cd764/OSTACK_IMG_be55a7c0-a3c7-4595-9d98-d55cc83cd764.vmdk {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 918.944525] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Downloading stream optimized image 8988d48c-9c55-4e0b-ad32-c37a4c0ae476 to [datastore2] OSTACK_IMG_be55a7c0-a3c7-4595-9d98-d55cc83cd764/OSTACK_IMG_be55a7c0-a3c7-4595-9d98-d55cc83cd764.vmdk on the data store datastore2 as vApp {{(pid=63372) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 918.944731] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Downloading image file data 8988d48c-9c55-4e0b-ad32-c37a4c0ae476 to the ESX as VM named 'OSTACK_IMG_be55a7c0-a3c7-4595-9d98-d55cc83cd764' {{(pid=63372) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 918.949689] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Acquiring lock "ed2cefa1-3604-43c5-9860-0fb49133bbb4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.950047] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Lock "ed2cefa1-3604-43c5-9860-0fb49133bbb4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.004581] env[63372]: DEBUG nova.objects.base [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Object Instance<19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5> lazy-loaded attributes: flavor,info_cache {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 919.031291] env[63372]: DEBUG oslo_vmware.rw_handles [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 919.031291] env[63372]: value = "resgroup-9" [ 919.031291] env[63372]: _type = "ResourcePool" [ 919.031291] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 919.031849] env[63372]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-d013ecf2-c5b0-404e-8d7e-38a5b6986573 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.052676] env[63372]: DEBUG oslo_vmware.rw_handles [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lease: (returnval){ [ 919.052676] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52171467-d271-f754-22aa-a39437d52f12" [ 919.052676] env[63372]: _type = "HttpNfcLease" [ 919.052676] env[63372]: } obtained for vApp import into resource pool (val){ [ 919.052676] env[63372]: value = "resgroup-9" [ 919.052676] env[63372]: _type = "ResourcePool" [ 919.052676] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 919.053356] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the lease: (returnval){ [ 919.053356] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52171467-d271-f754-22aa-a39437d52f12" [ 919.053356] env[63372]: _type = "HttpNfcLease" [ 919.053356] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 919.059500] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 919.059500] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52171467-d271-f754-22aa-a39437d52f12" [ 919.059500] env[63372]: _type = "HttpNfcLease" [ 919.059500] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 919.067657] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e45bcd03-e3eb-40e8-9508-9d83e7cf7cd0 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-13bde897-8446-42a2-b02d-2f5b48e6f432-eac444f7-19cf-4cc8-b354-21b08d90753c" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.024s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.272173] env[63372]: DEBUG oslo_vmware.api [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024425, 'name': PowerOnVM_Task, 'duration_secs': 0.755963} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.272666] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 919.272733] env[63372]: DEBUG nova.compute.manager [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.273492] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be01bb68-ea8b-4661-8a75-4c61ae6ada01 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.327499] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024426, 'name': CreateVM_Task, 'duration_secs': 0.527674} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.327659] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 919.329071] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.329295] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.330518] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 919.331317] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5eafd2b9-476c-4193-9146-14791acb706d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.337568] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 919.337568] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fdd529-64fc-ba61-5cee-54574fcf7cf5" [ 919.337568] env[63372]: _type = "Task" [ 919.337568] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.338638] env[63372]: DEBUG nova.scheduler.client.report [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.351505] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fdd529-64fc-ba61-5cee-54574fcf7cf5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.452280] env[63372]: DEBUG nova.compute.manager [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 919.561366] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 919.561366] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52171467-d271-f754-22aa-a39437d52f12" [ 919.561366] env[63372]: _type = "HttpNfcLease" [ 919.561366] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 919.792465] env[63372]: DEBUG oslo_concurrency.lockutils [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.849378] env[63372]: DEBUG oslo_concurrency.lockutils [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.849866] env[63372]: DEBUG nova.compute.manager [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 919.854798] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fdd529-64fc-ba61-5cee-54574fcf7cf5, 'name': SearchDatastore_Task, 'duration_secs': 0.018682} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.855458] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.369s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.856161] env[63372]: DEBUG nova.objects.instance [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lazy-loading 'resources' on Instance uuid 75560852-8c6d-4220-b251-81e8721a4b7b {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.856688] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.858021] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.858021] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.858169] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.858368] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.859165] env[63372]: DEBUG nova.network.neutron [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updated VIF entry in instance network info cache for port c92e8cc9-dc72-4f20-b087-1d323c502108. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 919.859525] env[63372]: DEBUG nova.network.neutron [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updating instance_info_cache with network_info: [{"id": "c92e8cc9-dc72-4f20-b087-1d323c502108", "address": "fa:16:3e:39:9e:be", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc92e8cc9-dc", "ovs_interfaceid": "c92e8cc9-dc72-4f20-b087-1d323c502108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.860949] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e9fa75e-fcd0-4a92-96d9-020ca100c2fe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.870433] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.870626] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 919.871424] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-657960a1-4d12-4558-88c9-b53e9be4243e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.879627] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 919.879627] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e8051b-8035-3553-c53d-efff473899fb" [ 919.879627] env[63372]: _type = "Task" [ 919.879627] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.889285] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e8051b-8035-3553-c53d-efff473899fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.950701] env[63372]: DEBUG nova.network.neutron [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance_info_cache with network_info: [{"id": "e1486286-9147-4e5e-88c1-60893a3338dd", "address": "fa:16:3e:6c:92:ac", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1486286-91", "ovs_interfaceid": "e1486286-9147-4e5e-88c1-60893a3338dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.978078] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.062542] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 920.062542] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52171467-d271-f754-22aa-a39437d52f12" [ 920.062542] env[63372]: _type = "HttpNfcLease" [ 920.062542] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 920.354917] env[63372]: DEBUG nova.compute.utils [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.356776] env[63372]: DEBUG nova.compute.manager [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Not allocating networking since 'none' was specified. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 920.361758] env[63372]: DEBUG oslo_concurrency.lockutils [req-a69a08aa-c490-42fe-bd8a-911f828cb140 req-81cb1d7a-eb13-4e1d-b812-b783ae9f8f8c service nova] Releasing lock "refresh_cache-c01a5d24-eb46-4a69-993e-753880ce8e85" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.390227] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e8051b-8035-3553-c53d-efff473899fb, 'name': SearchDatastore_Task, 'duration_secs': 0.01272} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.393346] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66a1abf8-5d4c-45a8-afa6-c442e8427584 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.398588] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 920.398588] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f2cc03-d8ab-3e96-8fbd-899a8d3cd498" [ 920.398588] env[63372]: _type = "Task" [ 920.398588] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.406183] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f2cc03-d8ab-3e96-8fbd-899a8d3cd498, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.457541] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "refresh_cache-19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.541955] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba5632b-06c6-4552-9d40-e9df10801ecc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.552031] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38e29e8-1569-4bcf-a8d8-ae678891b072 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.565457] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 920.565457] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52171467-d271-f754-22aa-a39437d52f12" [ 920.565457] env[63372]: _type = "HttpNfcLease" [ 920.565457] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 920.592579] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0530e26a-d8da-48bf-8597-0ead2f6519f5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.600640] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ea85b9-302f-45ca-8371-63fdff7bf4eb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.615945] env[63372]: DEBUG nova.compute.provider_tree [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.858183] env[63372]: DEBUG nova.compute.manager [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 920.911276] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f2cc03-d8ab-3e96-8fbd-899a8d3cd498, 'name': SearchDatastore_Task, 'duration_secs': 0.013992} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.911558] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.911859] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 326a9155-1c00-4c18-a9a7-6124e709c20b/326a9155-1c00-4c18-a9a7-6124e709c20b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 920.912188] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-edf8a223-1d60-4767-bd38-42b4e88ca863 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.921097] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 920.921097] env[63372]: value = "task-1024428" [ 920.921097] env[63372]: _type = "Task" [ 920.921097] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.929258] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024428, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.961503] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 920.961892] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8198941b-944b-4576-a4bf-e3008b49104b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.969390] env[63372]: DEBUG oslo_vmware.api [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 920.969390] env[63372]: value = "task-1024429" [ 920.969390] env[63372]: _type = "Task" [ 920.969390] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.977821] env[63372]: DEBUG oslo_vmware.api [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024429, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.033185] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "interface-13bde897-8446-42a2-b02d-2f5b48e6f432-a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.033476] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-13bde897-8446-42a2-b02d-2f5b48e6f432-a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.066954] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 921.066954] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52171467-d271-f754-22aa-a39437d52f12" [ 921.066954] env[63372]: _type = "HttpNfcLease" [ 921.066954] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 921.067512] env[63372]: DEBUG oslo_vmware.rw_handles [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 921.067512] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52171467-d271-f754-22aa-a39437d52f12" [ 921.067512] env[63372]: _type = "HttpNfcLease" [ 921.067512] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 921.068747] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9cff373-57cc-470b-9f7d-7629c48b0068 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.079878] env[63372]: DEBUG oslo_vmware.rw_handles [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5217ef60-599c-f1fa-33c9-701577d351bb/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 921.079878] env[63372]: DEBUG oslo_vmware.rw_handles [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Creating HTTP connection to write to file with size = 31662592 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5217ef60-599c-f1fa-33c9-701577d351bb/disk-0.vmdk. {{(pid=63372) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 921.146198] env[63372]: DEBUG nova.scheduler.client.report [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.158433] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e160bb69-5272-4247-81b4-939fdb6660f7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.288803] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "d568b727-7836-4b1d-9f27-f159227e46f7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.289215] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "d568b727-7836-4b1d-9f27-f159227e46f7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.289616] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "d568b727-7836-4b1d-9f27-f159227e46f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.289875] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "d568b727-7836-4b1d-9f27-f159227e46f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.290122] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "d568b727-7836-4b1d-9f27-f159227e46f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.293110] env[63372]: INFO nova.compute.manager [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Terminating instance [ 921.295770] env[63372]: DEBUG nova.compute.manager [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 921.295981] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 921.297214] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cdfbd96-56e7-4208-a76f-73325beee315 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.309028] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 921.309028] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3c252f29-ff27-4bff-9c64-5c5bd7b7600e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.315270] env[63372]: DEBUG oslo_vmware.api [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 921.315270] env[63372]: value = "task-1024430" [ 921.315270] env[63372]: _type = "Task" [ 921.315270] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.327717] env[63372]: DEBUG oslo_vmware.api [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024430, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.432030] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024428, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508326} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.432423] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 326a9155-1c00-4c18-a9a7-6124e709c20b/326a9155-1c00-4c18-a9a7-6124e709c20b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 921.432533] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 921.432768] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd2d2c70-0c84-4f8d-9522-197542d6f729 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.439740] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 921.439740] env[63372]: value = "task-1024431" [ 921.439740] env[63372]: _type = "Task" [ 921.439740] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.448455] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024431, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.478579] env[63372]: DEBUG oslo_vmware.api [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024429, 'name': PowerOnVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.535816] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.536028] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.537056] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed47d693-7b27-4b78-9c8b-a534c02559fa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.557132] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd8764b6-5b6b-4d4e-9269-c8e95fc3fd6e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.585267] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Reconfiguring VM to detach interface {{(pid=63372) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 921.588373] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac42b5b0-677e-450e-b43a-708c0bce8ebb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.609589] env[63372]: DEBUG oslo_vmware.api [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 921.609589] env[63372]: value = "task-1024432" [ 921.609589] env[63372]: _type = "Task" [ 921.609589] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.619623] env[63372]: DEBUG oslo_vmware.api [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024432, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.653494] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.798s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.655837] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.609s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.657358] env[63372]: INFO nova.compute.claims [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.684318] env[63372]: INFO nova.scheduler.client.report [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleted allocations for instance 75560852-8c6d-4220-b251-81e8721a4b7b [ 921.829904] env[63372]: DEBUG oslo_vmware.api [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024430, 'name': PowerOffVM_Task, 'duration_secs': 0.299576} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.832082] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 921.832082] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 921.832082] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ccf505da-222f-4f88-8cdb-e3efdc90a42f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.871069] env[63372]: DEBUG nova.compute.manager [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 921.901086] env[63372]: DEBUG nova.virt.hardware [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 921.901455] env[63372]: DEBUG nova.virt.hardware [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 921.901714] env[63372]: DEBUG nova.virt.hardware [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 921.902010] env[63372]: DEBUG nova.virt.hardware [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 921.902309] env[63372]: DEBUG nova.virt.hardware [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 921.902547] env[63372]: DEBUG nova.virt.hardware [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 921.902811] env[63372]: DEBUG nova.virt.hardware [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 921.902994] env[63372]: DEBUG nova.virt.hardware [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 921.903183] env[63372]: DEBUG nova.virt.hardware [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 921.903376] env[63372]: DEBUG nova.virt.hardware [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 921.903589] env[63372]: DEBUG nova.virt.hardware [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 921.904535] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a94bc70-71c3-4381-939a-2d1afb0a56d3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.913138] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26dfce4a-8e68-4f0e-b156-e03f7854e951 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.921554] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 921.922063] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 921.922415] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleting the datastore file [datastore2] d568b727-7836-4b1d-9f27-f159227e46f7 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.931216] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4dade98e-599d-4775-a75e-cb0e2cd7a447 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.933186] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Instance VIF info [] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.938948] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Creating folder: Project (822033fd2ab9404c9e39e1f9bce543a8). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 921.939397] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cf100b5c-6f6e-4e2c-9176-040acb8f601a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.946235] env[63372]: DEBUG oslo_vmware.api [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 921.946235] env[63372]: value = "task-1024434" [ 921.946235] env[63372]: _type = "Task" [ 921.946235] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.953184] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024431, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064489} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.954748] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 921.955149] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Created folder: Project (822033fd2ab9404c9e39e1f9bce543a8) in parent group-v227230. [ 921.955382] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Creating folder: Instances. Parent ref: group-v227410. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 921.956200] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17e48cf-eb6c-412a-8b66-5a69f0b87dac {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.958744] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-666d079f-e7d6-47ff-9499-a15883efd40e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.963760] env[63372]: DEBUG oslo_vmware.api [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024434, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.986465] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 326a9155-1c00-4c18-a9a7-6124e709c20b/326a9155-1c00-4c18-a9a7-6124e709c20b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.996303] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67c701d5-27cc-4f1e-a759-957a10bd98ef {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.012099] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Created folder: Instances in parent group-v227410. [ 922.012451] env[63372]: DEBUG oslo.service.loopingcall [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.017414] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 922.018352] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5dc7a410-8f2e-4929-aea1-52ac5aad8480 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.039337] env[63372]: DEBUG oslo_vmware.api [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024429, 'name': PowerOnVM_Task, 'duration_secs': 0.559848} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.039777] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 922.039777] env[63372]: value = "task-1024437" [ 922.039777] env[63372]: _type = "Task" [ 922.039777] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.040615] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 922.040930] env[63372]: DEBUG nova.compute.manager [None req-1026e912-b628-4293-8a46-cbfefd739488 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.041908] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c5ecd2-af03-4548-bc50-30992aa7dde2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.051100] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.051100] env[63372]: value = "task-1024438" [ 922.051100] env[63372]: _type = "Task" [ 922.051100] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.061189] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024437, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.073148] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024438, 'name': CreateVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.125431] env[63372]: DEBUG oslo_vmware.api [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.195801] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2a1b9b30-e6ac-42cd-90a3-0edcd22de52b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "75560852-8c6d-4220-b251-81e8721a4b7b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.951s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.460150] env[63372]: DEBUG oslo_vmware.api [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024434, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166406} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.461566] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.461773] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 922.461994] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 922.462288] env[63372]: INFO nova.compute.manager [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Took 1.17 seconds to destroy the instance on the hypervisor. [ 922.462571] env[63372]: DEBUG oslo.service.loopingcall [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.462998] env[63372]: DEBUG nova.compute.manager [-] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 922.462998] env[63372]: DEBUG nova.network.neutron [-] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 922.555527] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024437, 'name': ReconfigVM_Task, 'duration_secs': 0.395388} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.560367] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 326a9155-1c00-4c18-a9a7-6124e709c20b/326a9155-1c00-4c18-a9a7-6124e709c20b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 922.561201] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a1881783-f8d0-42a4-b1f5-5f1affa51bba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.569340] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024438, 'name': CreateVM_Task, 'duration_secs': 0.329302} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.571874] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 922.572239] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 922.572239] env[63372]: value = "task-1024439" [ 922.572239] env[63372]: _type = "Task" [ 922.572239] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.573529] env[63372]: DEBUG oslo_concurrency.lockutils [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.573683] env[63372]: DEBUG oslo_concurrency.lockutils [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.574048] env[63372]: DEBUG oslo_concurrency.lockutils [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 922.574505] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c44aeb09-6900-4ca6-ae5f-8cb68f1c4d72 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.585827] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024439, 'name': Rename_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.586179] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 922.586179] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529361ca-46c9-0752-f7a9-6f956b0f81f7" [ 922.586179] env[63372]: _type = "Task" [ 922.586179] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.596984] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529361ca-46c9-0752-f7a9-6f956b0f81f7, 'name': SearchDatastore_Task, 'duration_secs': 0.011482} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.598631] env[63372]: DEBUG oslo_concurrency.lockutils [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.598904] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.599325] env[63372]: DEBUG oslo_concurrency.lockutils [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.599422] env[63372]: DEBUG oslo_concurrency.lockutils [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.599926] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.599926] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-66c94fb8-aead-45f4-9ac4-125196801c5b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.607538] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.607805] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 922.608528] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3150f341-8b1d-4d44-aa96-1eceb89d0e7a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.618747] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 922.618747] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52c635c7-1168-b897-604b-bccb33419b92" [ 922.618747] env[63372]: _type = "Task" [ 922.618747] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.627166] env[63372]: DEBUG oslo_vmware.api [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.632528] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c635c7-1168-b897-604b-bccb33419b92, 'name': SearchDatastore_Task, 'duration_secs': 0.008952} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.633357] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a83db79-153a-4208-8522-d1a517db9c95 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.638506] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 922.638506] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]524cecb1-97ee-69a3-f77c-45eb3551771b" [ 922.638506] env[63372]: _type = "Task" [ 922.638506] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.647738] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524cecb1-97ee-69a3-f77c-45eb3551771b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.663446] env[63372]: DEBUG oslo_vmware.rw_handles [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Completed reading data from the image iterator. {{(pid=63372) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 922.663767] env[63372]: DEBUG oslo_vmware.rw_handles [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5217ef60-599c-f1fa-33c9-701577d351bb/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 922.664866] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6094de-6ee8-4a6d-92da-ad16a5966ab0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.670588] env[63372]: DEBUG oslo_vmware.rw_handles [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5217ef60-599c-f1fa-33c9-701577d351bb/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 922.670853] env[63372]: DEBUG oslo_vmware.rw_handles [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5217ef60-599c-f1fa-33c9-701577d351bb/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 922.670987] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-1eed4d20-e9ad-40c7-ad4b-d77fe67c3876 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.891822] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49fbdef7-12aa-450d-9476-135981f93e40 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.905312] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b29798-3407-447f-b96a-5bb128fc2d46 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.944701] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2436cdaa-8433-48e3-8a04-beb740cbfbb8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.953829] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce6e2fd-fb74-4050-8609-210217602ac0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.969895] env[63372]: DEBUG nova.compute.provider_tree [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.973089] env[63372]: DEBUG nova.compute.manager [req-46ca84de-28a3-465c-967a-9552a71d4151 req-4d82075e-ac61-4cd0-818d-3f38964bb26f service nova] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Received event network-vif-deleted-2454845a-285a-4d06-8f95-123e004fae02 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.973330] env[63372]: INFO nova.compute.manager [req-46ca84de-28a3-465c-967a-9552a71d4151 req-4d82075e-ac61-4cd0-818d-3f38964bb26f service nova] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Neutron deleted interface 2454845a-285a-4d06-8f95-123e004fae02; detaching it from the instance and deleting it from the info cache [ 922.973543] env[63372]: DEBUG nova.network.neutron [req-46ca84de-28a3-465c-967a-9552a71d4151 req-4d82075e-ac61-4cd0-818d-3f38964bb26f service nova] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.079657] env[63372]: DEBUG oslo_vmware.rw_handles [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5217ef60-599c-f1fa-33c9-701577d351bb/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 923.079963] env[63372]: INFO nova.virt.vmwareapi.images [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Downloaded image file data 8988d48c-9c55-4e0b-ad32-c37a4c0ae476 [ 923.081186] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61cecdba-355e-45e5-994c-d06ab8edc116 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.087330] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024439, 'name': Rename_Task, 'duration_secs': 0.167643} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.088038] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 923.088330] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-adaabd18-21c5-47da-ac99-51b4b7912823 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.100214] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-711f27bd-907f-46e9-a89f-0fc953f0b05a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.104287] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 923.104287] env[63372]: value = "task-1024440" [ 923.104287] env[63372]: _type = "Task" [ 923.104287] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.112426] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024440, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.122229] env[63372]: DEBUG oslo_vmware.api [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.148602] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524cecb1-97ee-69a3-f77c-45eb3551771b, 'name': SearchDatastore_Task, 'duration_secs': 0.010268} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.148884] env[63372]: DEBUG oslo_concurrency.lockutils [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.149173] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 662d9fdf-ec0d-4466-8d15-abc942e4dad8/662d9fdf-ec0d-4466-8d15-abc942e4dad8.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 923.149471] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-48f9f87d-788c-457e-b85d-d790363d597d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.155624] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 923.155624] env[63372]: value = "task-1024442" [ 923.155624] env[63372]: _type = "Task" [ 923.155624] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.163953] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024442, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.192517] env[63372]: INFO nova.virt.vmwareapi.images [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] The imported VM was unregistered [ 923.195679] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Caching image {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 923.195943] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Creating directory with path [datastore2] devstack-image-cache_base/8988d48c-9c55-4e0b-ad32-c37a4c0ae476 {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.196287] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-79a1cfc3-d1e1-4cee-b76c-8081a14811cb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.230306] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Created directory with path [datastore2] devstack-image-cache_base/8988d48c-9c55-4e0b-ad32-c37a4c0ae476 {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.230601] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_be55a7c0-a3c7-4595-9d98-d55cc83cd764/OSTACK_IMG_be55a7c0-a3c7-4595-9d98-d55cc83cd764.vmdk to [datastore2] devstack-image-cache_base/8988d48c-9c55-4e0b-ad32-c37a4c0ae476/8988d48c-9c55-4e0b-ad32-c37a4c0ae476.vmdk. {{(pid=63372) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 923.230887] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-0d692edc-27a9-40af-8e8c-4c33c8b4f8a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.239027] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 923.239027] env[63372]: value = "task-1024443" [ 923.239027] env[63372]: _type = "Task" [ 923.239027] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.247073] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024443, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.399097] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.399617] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.399905] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.400193] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.400472] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.403419] env[63372]: INFO nova.compute.manager [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Terminating instance [ 923.405746] env[63372]: DEBUG nova.compute.manager [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 923.406082] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 923.407462] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00725d2d-f9fe-4632-a678-6aa33017688b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.421533] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 923.421930] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5aa6e368-bcbc-4a88-94cc-f098104c6eca {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.431256] env[63372]: DEBUG oslo_vmware.api [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 923.431256] env[63372]: value = "task-1024444" [ 923.431256] env[63372]: _type = "Task" [ 923.431256] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.441321] env[63372]: DEBUG oslo_vmware.api [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024444, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.448395] env[63372]: DEBUG nova.network.neutron [-] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.472930] env[63372]: DEBUG nova.scheduler.client.report [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.476850] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a75942e8-bce5-4940-b298-933176bffb77 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.483814] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.484125] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.495284] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45258ffc-ac6e-44e6-a7cf-010aee73c49d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.532449] env[63372]: DEBUG nova.compute.manager [req-46ca84de-28a3-465c-967a-9552a71d4151 req-4d82075e-ac61-4cd0-818d-3f38964bb26f service nova] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Detach interface failed, port_id=2454845a-285a-4d06-8f95-123e004fae02, reason: Instance d568b727-7836-4b1d-9f27-f159227e46f7 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 923.615914] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024440, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.625134] env[63372]: DEBUG oslo_vmware.api [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.668053] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024442, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.754580] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024443, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.943708] env[63372]: DEBUG oslo_vmware.api [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024444, 'name': PowerOffVM_Task, 'duration_secs': 0.239701} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.944068] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 923.944302] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 923.944588] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56c04be1-4cf3-49b4-a8a7-b0cdba0e2ce9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.951173] env[63372]: INFO nova.compute.manager [-] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Took 1.49 seconds to deallocate network for instance. [ 923.977782] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.981143] env[63372]: DEBUG nova.compute.manager [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 923.981318] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.540s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.988031] env[63372]: DEBUG nova.compute.manager [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 924.027401] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 924.027816] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 924.028183] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleting the datastore file [datastore2] 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.028607] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9a83cbb4-97de-4354-8880-44d501c32c08 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.039562] env[63372]: DEBUG oslo_vmware.api [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 924.039562] env[63372]: value = "task-1024446" [ 924.039562] env[63372]: _type = "Task" [ 924.039562] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.051793] env[63372]: DEBUG oslo_vmware.api [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024446, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.117336] env[63372]: DEBUG oslo_vmware.api [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024440, 'name': PowerOnVM_Task, 'duration_secs': 0.741414} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.121159] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 924.121455] env[63372]: DEBUG nova.compute.manager [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.122223] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6bfa26f-5c4d-4c13-9261-bdc404a1ab3e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.138117] env[63372]: DEBUG oslo_vmware.api [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.171028] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024442, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.887036} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.171401] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 662d9fdf-ec0d-4466-8d15-abc942e4dad8/662d9fdf-ec0d-4466-8d15-abc942e4dad8.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 924.171730] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 924.172302] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-76f0bed0-7858-4719-8753-11f26dc82145 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.182203] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 924.182203] env[63372]: value = "task-1024447" [ 924.182203] env[63372]: _type = "Task" [ 924.182203] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.193613] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024447, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.252491] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024443, 'name': MoveVirtualDisk_Task} progress is 29%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.459471] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.485222] env[63372]: DEBUG nova.compute.utils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 924.486977] env[63372]: DEBUG nova.compute.manager [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 924.487214] env[63372]: DEBUG nova.network.neutron [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 924.518895] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.554894] env[63372]: DEBUG oslo_vmware.api [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024446, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.629440] env[63372]: DEBUG oslo_vmware.api [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.645784] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.695217] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024447, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074654} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.695567] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 924.696542] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575ceb27-40f0-43bd-bd75-63a3e00a763f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.721336] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 662d9fdf-ec0d-4466-8d15-abc942e4dad8/662d9fdf-ec0d-4466-8d15-abc942e4dad8.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 924.721713] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a42f4da-29d0-4abd-95a4-298a9e5bd553 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.746401] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 924.746401] env[63372]: value = "task-1024448" [ 924.746401] env[63372]: _type = "Task" [ 924.746401] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.754128] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024443, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.759700] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.839890] env[63372]: DEBUG nova.policy [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb233d4b6fe54fefae3046d62d167e34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d0c0f315749429dbae22a72ae82b1b2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 924.990786] env[63372]: DEBUG nova.compute.manager [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 925.020783] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance c4718797-aa86-4ec0-94d3-6480bd6aa898 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.020977] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 93a5d948-0629-4f53-a681-858d519acfa7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.021146] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 13bde897-8446-42a2-b02d-2f5b48e6f432 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.021467] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.021755] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance c01a5d24-eb46-4a69-993e-753880ce8e85 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.021755] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance ba1daa0a-dcf8-4586-9964-a962fcc94196 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.022023] env[63372]: WARNING nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance d568b727-7836-4b1d-9f27-f159227e46f7 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 925.022023] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 326a9155-1c00-4c18-a9a7-6124e709c20b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.022248] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 662d9fdf-ec0d-4466-8d15-abc942e4dad8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.022248] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 95178e2e-9d71-4606-a856-c530bfbb9345 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.057587] env[63372]: DEBUG oslo_vmware.api [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024446, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.132881] env[63372]: DEBUG oslo_vmware.api [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.256948] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024443, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.263708] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.526792] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 91985614-b959-401e-bb06-d67b230ee026 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 925.558614] env[63372]: DEBUG oslo_vmware.api [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024446, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.635204] env[63372]: DEBUG oslo_vmware.api [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.646686] env[63372]: DEBUG nova.network.neutron [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Successfully created port: 974ab817-fe7c-424a-9174-137c3ee524f4 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 925.758868] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024443, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.767973] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.002776] env[63372]: DEBUG nova.compute.manager [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 926.040146] env[63372]: DEBUG nova.virt.hardware [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 926.040631] env[63372]: DEBUG nova.virt.hardware [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 926.040884] env[63372]: DEBUG nova.virt.hardware [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.041233] env[63372]: DEBUG nova.virt.hardware [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 926.041498] env[63372]: DEBUG nova.virt.hardware [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.041759] env[63372]: DEBUG nova.virt.hardware [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 926.042129] env[63372]: DEBUG nova.virt.hardware [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 926.042408] env[63372]: DEBUG nova.virt.hardware [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 926.042703] env[63372]: DEBUG nova.virt.hardware [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 926.042999] env[63372]: DEBUG nova.virt.hardware [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 926.043366] env[63372]: DEBUG nova.virt.hardware [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.044706] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance ed2cefa1-3604-43c5-9860-0fb49133bbb4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 926.047239] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ace882-2f70-48a0-b8e2-bd27ed39dcca {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.068809] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4182e6-6985-4374-8df2-6d7c39050c17 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.076857] env[63372]: DEBUG oslo_vmware.api [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024446, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.822024} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.077361] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.077648] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 926.077930] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 926.078288] env[63372]: INFO nova.compute.manager [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Took 2.67 seconds to destroy the instance on the hypervisor. [ 926.078693] env[63372]: DEBUG oslo.service.loopingcall [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.079876] env[63372]: DEBUG nova.compute.manager [-] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 926.080033] env[63372]: DEBUG nova.network.neutron [-] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 926.132034] env[63372]: DEBUG oslo_vmware.api [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.251344] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024443, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.259049] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.334820] env[63372]: DEBUG oslo_concurrency.lockutils [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "326a9155-1c00-4c18-a9a7-6124e709c20b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.335116] env[63372]: DEBUG oslo_concurrency.lockutils [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "326a9155-1c00-4c18-a9a7-6124e709c20b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.335342] env[63372]: DEBUG oslo_concurrency.lockutils [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "326a9155-1c00-4c18-a9a7-6124e709c20b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.335544] env[63372]: DEBUG oslo_concurrency.lockutils [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "326a9155-1c00-4c18-a9a7-6124e709c20b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.335726] env[63372]: DEBUG oslo_concurrency.lockutils [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "326a9155-1c00-4c18-a9a7-6124e709c20b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.339275] env[63372]: INFO nova.compute.manager [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Terminating instance [ 926.342035] env[63372]: DEBUG nova.compute.manager [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 926.342320] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 926.343135] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a67018-625d-478a-8285-1c35a0fda97c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.353465] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 926.353728] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8530f794-fd0c-4218-b468-dff586787ede {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.361938] env[63372]: DEBUG oslo_vmware.api [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 926.361938] env[63372]: value = "task-1024449" [ 926.361938] env[63372]: _type = "Task" [ 926.361938] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.372143] env[63372]: DEBUG oslo_vmware.api [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024449, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.553983] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 926.554302] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 926.554403] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2304MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 926.636929] env[63372]: DEBUG oslo_vmware.api [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.755460] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024443, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.147257} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.759398] env[63372]: INFO nova.virt.vmwareapi.ds_util [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_be55a7c0-a3c7-4595-9d98-d55cc83cd764/OSTACK_IMG_be55a7c0-a3c7-4595-9d98-d55cc83cd764.vmdk to [datastore2] devstack-image-cache_base/8988d48c-9c55-4e0b-ad32-c37a4c0ae476/8988d48c-9c55-4e0b-ad32-c37a4c0ae476.vmdk. [ 926.760063] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Cleaning up location [datastore2] OSTACK_IMG_be55a7c0-a3c7-4595-9d98-d55cc83cd764 {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 926.760063] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_be55a7c0-a3c7-4595-9d98-d55cc83cd764 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 926.760300] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb6bf797-9cad-42c3-a234-3ed72b2cd6c9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.770436] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024448, 'name': ReconfigVM_Task, 'duration_secs': 1.827157} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.772935] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 662d9fdf-ec0d-4466-8d15-abc942e4dad8/662d9fdf-ec0d-4466-8d15-abc942e4dad8.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.774106] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 926.774106] env[63372]: value = "task-1024450" [ 926.774106] env[63372]: _type = "Task" [ 926.774106] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.774106] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-49ea41c5-4747-4e54-ab93-7cb29442b05d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.788758] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024450, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.790553] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 926.790553] env[63372]: value = "task-1024451" [ 926.790553] env[63372]: _type = "Task" [ 926.790553] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.798783] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024451, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.800315] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0f536a-eb82-4779-a292-5e8e9592fffa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.806990] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0b7257-1c85-4784-95cb-4cf555335795 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.846195] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279f9a60-ff82-48d5-814f-865f8e13881f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.853952] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ea14d7-938b-438f-bd71-eff85c181446 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.868610] env[63372]: DEBUG nova.compute.provider_tree [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.879147] env[63372]: DEBUG oslo_vmware.api [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024449, 'name': PowerOffVM_Task, 'duration_secs': 0.27436} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.879514] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 926.879840] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 926.880691] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2beac84d-e9aa-42b8-bddf-990d420cd2a0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.901136] env[63372]: DEBUG nova.compute.manager [req-06522e4a-215e-4332-b7aa-b176edadfe62 req-6c66f670-76b9-47c1-8301-9465813f8606 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Received event network-vif-deleted-e1486286-9147-4e5e-88c1-60893a3338dd {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.901136] env[63372]: INFO nova.compute.manager [req-06522e4a-215e-4332-b7aa-b176edadfe62 req-6c66f670-76b9-47c1-8301-9465813f8606 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Neutron deleted interface e1486286-9147-4e5e-88c1-60893a3338dd; detaching it from the instance and deleting it from the info cache [ 926.901136] env[63372]: DEBUG nova.network.neutron [req-06522e4a-215e-4332-b7aa-b176edadfe62 req-6c66f670-76b9-47c1-8301-9465813f8606 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.973607] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 926.973607] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 926.973813] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleting the datastore file [datastore2] 326a9155-1c00-4c18-a9a7-6124e709c20b {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 926.974067] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b4e6818f-d35e-4c26-9f9a-b9959961d875 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.980922] env[63372]: DEBUG oslo_vmware.api [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 926.980922] env[63372]: value = "task-1024453" [ 926.980922] env[63372]: _type = "Task" [ 926.980922] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.988335] env[63372]: DEBUG oslo_vmware.api [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024453, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.131948] env[63372]: DEBUG oslo_vmware.api [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024432, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.163076] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "61f7a052-9a42-4df3-a568-eab197fc5c4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.163323] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "61f7a052-9a42-4df3-a568-eab197fc5c4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.285235] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024450, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.044603} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.285506] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.285677] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8988d48c-9c55-4e0b-ad32-c37a4c0ae476/8988d48c-9c55-4e0b-ad32-c37a4c0ae476.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.286421] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8988d48c-9c55-4e0b-ad32-c37a4c0ae476/8988d48c-9c55-4e0b-ad32-c37a4c0ae476.vmdk to [datastore2] c01a5d24-eb46-4a69-993e-753880ce8e85/c01a5d24-eb46-4a69-993e-753880ce8e85.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 927.286421] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c6ff91cc-1128-4287-bf98-71642d1bda47 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.294624] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 927.294624] env[63372]: value = "task-1024454" [ 927.294624] env[63372]: _type = "Task" [ 927.294624] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.301689] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024451, 'name': Rename_Task, 'duration_secs': 0.143663} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.302271] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 927.302539] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6fdfd49-a040-4286-aef3-0c1b1f7132aa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.307100] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024454, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.311326] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 927.311326] env[63372]: value = "task-1024455" [ 927.311326] env[63372]: _type = "Task" [ 927.311326] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.318777] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024455, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.376202] env[63372]: DEBUG nova.scheduler.client.report [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.379579] env[63372]: DEBUG nova.network.neutron [-] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.405350] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a11d8a2-0195-48f6-b859-ef8adce56cb7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.424753] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea203d3e-69a7-4474-a36c-129ceb41fbe1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.264593] env[63372]: DEBUG nova.network.neutron [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Successfully updated port: 974ab817-fe7c-424a-9174-137c3ee524f4 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 928.266292] env[63372]: DEBUG nova.compute.manager [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 928.269210] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 928.269513] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.288s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.269881] env[63372]: INFO nova.compute.manager [-] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Took 2.19 seconds to deallocate network for instance. [ 928.285536] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.700s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.287021] env[63372]: INFO nova.compute.claims [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 928.291562] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Acquiring lock "c7c444cb-0aaa-4486-840a-e48fa255b421" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.291823] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Lock "c7c444cb-0aaa-4486-840a-e48fa255b421" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.302684] env[63372]: DEBUG oslo_vmware.api [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024432, 'name': ReconfigVM_Task, 'duration_secs': 5.751679} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.319520] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.319775] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Reconfigured VM to detach interface {{(pid=63372) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 928.323771] env[63372]: DEBUG nova.compute.manager [req-06522e4a-215e-4332-b7aa-b176edadfe62 req-6c66f670-76b9-47c1-8301-9465813f8606 service nova] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Detach interface failed, port_id=e1486286-9147-4e5e-88c1-60893a3338dd, reason: Instance 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 928.324578] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024455, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.324842] env[63372]: DEBUG oslo_vmware.api [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024453, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161382} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.325082] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024454, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.325616] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.325809] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 928.326012] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 928.326190] env[63372]: INFO nova.compute.manager [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Took 1.98 seconds to destroy the instance on the hypervisor. [ 928.326491] env[63372]: DEBUG oslo.service.loopingcall [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.327097] env[63372]: DEBUG nova.compute.manager [-] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 928.327192] env[63372]: DEBUG nova.network.neutron [-] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 928.779562] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "refresh_cache-95178e2e-9d71-4606-a856-c530bfbb9345" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.780484] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "refresh_cache-95178e2e-9d71-4606-a856-c530bfbb9345" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.780484] env[63372]: DEBUG nova.network.neutron [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 928.792687] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024454, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.797793] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024455, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.799764] env[63372]: DEBUG nova.compute.manager [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 928.803811] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.816756] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.941596] env[63372]: DEBUG nova.compute.manager [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Received event network-vif-plugged-974ab817-fe7c-424a-9174-137c3ee524f4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.941596] env[63372]: DEBUG oslo_concurrency.lockutils [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] Acquiring lock "95178e2e-9d71-4606-a856-c530bfbb9345-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.941739] env[63372]: DEBUG oslo_concurrency.lockutils [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] Lock "95178e2e-9d71-4606-a856-c530bfbb9345-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.942879] env[63372]: DEBUG oslo_concurrency.lockutils [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] Lock "95178e2e-9d71-4606-a856-c530bfbb9345-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.942879] env[63372]: DEBUG nova.compute.manager [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] No waiting events found dispatching network-vif-plugged-974ab817-fe7c-424a-9174-137c3ee524f4 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 928.942879] env[63372]: WARNING nova.compute.manager [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Received unexpected event network-vif-plugged-974ab817-fe7c-424a-9174-137c3ee524f4 for instance with vm_state building and task_state spawning. [ 928.942879] env[63372]: DEBUG nova.compute.manager [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Received event network-changed-974ab817-fe7c-424a-9174-137c3ee524f4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.942879] env[63372]: DEBUG nova.compute.manager [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Refreshing instance network info cache due to event network-changed-974ab817-fe7c-424a-9174-137c3ee524f4. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 928.942879] env[63372]: DEBUG oslo_concurrency.lockutils [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] Acquiring lock "refresh_cache-95178e2e-9d71-4606-a856-c530bfbb9345" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.284177] env[63372]: DEBUG oslo_vmware.api [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024455, 'name': PowerOnVM_Task, 'duration_secs': 1.568353} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.287489] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 929.287732] env[63372]: INFO nova.compute.manager [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Took 7.42 seconds to spawn the instance on the hypervisor. [ 929.287923] env[63372]: DEBUG nova.compute.manager [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.290087] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024454, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.290936] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2bebf5-f359-417f-9605-fa47ee6db672 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.320628] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.339870] env[63372]: DEBUG nova.network.neutron [-] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.397180] env[63372]: DEBUG nova.network.neutron [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 929.536567] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866a9ab7-b615-4f1c-ada4-1f7124794231 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.546878] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421f7f6d-9451-47ff-8529-3254e0035abf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.580897] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab651ec-0a7c-4950-b268-dc842a3afd89 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.591561] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ccc2b7f-d0bf-4487-8a41-3e356bf282b9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.610840] env[63372]: DEBUG nova.compute.provider_tree [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.635787] env[63372]: DEBUG nova.network.neutron [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Updating instance_info_cache with network_info: [{"id": "974ab817-fe7c-424a-9174-137c3ee524f4", "address": "fa:16:3e:a6:5b:85", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap974ab817-fe", "ovs_interfaceid": "974ab817-fe7c-424a-9174-137c3ee524f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.726088] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "13bde897-8446-42a2-b02d-2f5b48e6f432" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.726088] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "13bde897-8446-42a2-b02d-2f5b48e6f432" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.726088] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "13bde897-8446-42a2-b02d-2f5b48e6f432-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.726088] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "13bde897-8446-42a2-b02d-2f5b48e6f432-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.726406] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "13bde897-8446-42a2-b02d-2f5b48e6f432-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.728670] env[63372]: INFO nova.compute.manager [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Terminating instance [ 929.730856] env[63372]: DEBUG nova.compute.manager [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 929.731082] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 929.731988] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b410afee-3e73-4b74-8234-bf69aacfdfd1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.741657] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 929.741956] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69672f74-3485-438a-8070-0195384d657f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.750747] env[63372]: DEBUG oslo_vmware.api [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 929.750747] env[63372]: value = "task-1024456" [ 929.750747] env[63372]: _type = "Task" [ 929.750747] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.759119] env[63372]: DEBUG oslo_vmware.api [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024456, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.779366] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024454, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.814718] env[63372]: INFO nova.compute.manager [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Took 23.01 seconds to build instance. [ 929.825879] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.826653] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.826653] env[63372]: DEBUG nova.network.neutron [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 929.847450] env[63372]: INFO nova.compute.manager [-] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Took 1.52 seconds to deallocate network for instance. [ 930.113975] env[63372]: DEBUG nova.scheduler.client.report [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.138020] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "refresh_cache-95178e2e-9d71-4606-a856-c530bfbb9345" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.138370] env[63372]: DEBUG nova.compute.manager [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Instance network_info: |[{"id": "974ab817-fe7c-424a-9174-137c3ee524f4", "address": "fa:16:3e:a6:5b:85", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap974ab817-fe", "ovs_interfaceid": "974ab817-fe7c-424a-9174-137c3ee524f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 930.139527] env[63372]: DEBUG oslo_concurrency.lockutils [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] Acquired lock "refresh_cache-95178e2e-9d71-4606-a856-c530bfbb9345" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.139527] env[63372]: DEBUG nova.network.neutron [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Refreshing network info cache for port 974ab817-fe7c-424a-9174-137c3ee524f4 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 930.140524] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a6:5b:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60badc2d-69d2-467d-a92e-98511f5cb0b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '974ab817-fe7c-424a-9174-137c3ee524f4', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.147887] env[63372]: DEBUG oslo.service.loopingcall [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.151252] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 930.151746] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58eae906-6da6-478f-907a-e00fe1e37d89 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.171839] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 930.171839] env[63372]: value = "task-1024457" [ 930.171839] env[63372]: _type = "Task" [ 930.171839] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.179536] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024457, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.263993] env[63372]: DEBUG oslo_vmware.api [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024456, 'name': PowerOffVM_Task, 'duration_secs': 0.227301} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.264349] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 930.264695] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 930.265065] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e47dac1e-7319-49b4-8bba-035a2abbda18 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.281291] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024454, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.316236] env[63372]: DEBUG oslo_concurrency.lockutils [None req-055a129e-4f20-4a0e-add1-f4adf3762820 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Lock "662d9fdf-ec0d-4466-8d15-abc942e4dad8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.523s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.355283] env[63372]: DEBUG oslo_concurrency.lockutils [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.508357] env[63372]: INFO nova.compute.manager [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Rebuilding instance [ 930.563660] env[63372]: DEBUG nova.compute.manager [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 930.564568] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e97c5e1-7fc5-44c6-b438-2b4b3cbab85a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.618756] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.619305] env[63372]: DEBUG nova.compute.manager [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 930.622060] env[63372]: DEBUG oslo_concurrency.lockutils [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 10.830s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.622173] env[63372]: DEBUG nova.objects.instance [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63372) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 930.676703] env[63372]: DEBUG nova.network.neutron [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Updated VIF entry in instance network info cache for port 974ab817-fe7c-424a-9174-137c3ee524f4. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 930.677059] env[63372]: DEBUG nova.network.neutron [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Updating instance_info_cache with network_info: [{"id": "974ab817-fe7c-424a-9174-137c3ee524f4", "address": "fa:16:3e:a6:5b:85", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap974ab817-fe", "ovs_interfaceid": "974ab817-fe7c-424a-9174-137c3ee524f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.683732] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024457, 'name': CreateVM_Task, 'duration_secs': 0.357647} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.683882] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 930.684550] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.684716] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.685089] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 930.685773] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5b7ba45-a739-4a40-892f-6cbad446a777 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.691121] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 930.691121] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]527634a3-9292-d002-e157-34e332da7d97" [ 930.691121] env[63372]: _type = "Task" [ 930.691121] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.698524] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527634a3-9292-d002-e157-34e332da7d97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.780779] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024454, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.801479] env[63372]: INFO nova.network.neutron [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Port a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 930.801585] env[63372]: INFO nova.network.neutron [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Port eac444f7-19cf-4cc8-b354-21b08d90753c from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 930.801929] env[63372]: DEBUG nova.network.neutron [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Updating instance_info_cache with network_info: [{"id": "b02b2fb0-9262-4f24-a677-c4f61362410a", "address": "fa:16:3e:03:c1:fa", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb02b2fb0-92", "ovs_interfaceid": "b02b2fb0-9262-4f24-a677-c4f61362410a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.966063] env[63372]: DEBUG nova.compute.manager [req-046afda0-b7ed-4c21-b487-8fab972a59ba req-c1953d34-b597-4e15-8dc3-d9be6fde0a53 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Received event network-vif-deleted-eac444f7-19cf-4cc8-b354-21b08d90753c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.076182] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 931.076486] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5d37cae6-8c0b-4495-9dce-d7f468297972 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.085938] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 931.085938] env[63372]: value = "task-1024459" [ 931.085938] env[63372]: _type = "Task" [ 931.085938] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.095283] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024459, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.126888] env[63372]: DEBUG nova.compute.utils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 931.132026] env[63372]: DEBUG nova.compute.manager [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 931.132026] env[63372]: DEBUG nova.network.neutron [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 931.172305] env[63372]: DEBUG nova.policy [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9592bafdf42843c18cb71d97f9fbd60a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e517915d5e7a4e3e924cc95c55170a66', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 931.180156] env[63372]: DEBUG oslo_concurrency.lockutils [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] Releasing lock "refresh_cache-95178e2e-9d71-4606-a856-c530bfbb9345" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.180449] env[63372]: DEBUG nova.compute.manager [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Received event network-vif-deleted-a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.181340] env[63372]: INFO nova.compute.manager [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Neutron deleted interface a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c; detaching it from the instance and deleting it from the info cache [ 931.181340] env[63372]: DEBUG nova.network.neutron [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Updating instance_info_cache with network_info: [{"id": "b02b2fb0-9262-4f24-a677-c4f61362410a", "address": "fa:16:3e:03:c1:fa", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb02b2fb0-92", "ovs_interfaceid": "b02b2fb0-9262-4f24-a677-c4f61362410a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "eac444f7-19cf-4cc8-b354-21b08d90753c", "address": "fa:16:3e:1a:e3:dd", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeac444f7-19", "ovs_interfaceid": "eac444f7-19cf-4cc8-b354-21b08d90753c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.201663] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527634a3-9292-d002-e157-34e332da7d97, 'name': SearchDatastore_Task, 'duration_secs': 0.05287} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.201965] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.202252] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 931.202493] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.202784] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.202879] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.203087] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9c568940-cf59-4ed9-b34a-32146a29062d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.212028] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.212227] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 931.212954] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-782f5bf6-a7a1-4707-b119-a5e9590c8a20 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.218820] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 931.218820] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52750143-00a7-33ad-da65-93fc9a461593" [ 931.218820] env[63372]: _type = "Task" [ 931.218820] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.227086] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52750143-00a7-33ad-da65-93fc9a461593, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.282876] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024454, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.515606} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.283171] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8988d48c-9c55-4e0b-ad32-c37a4c0ae476/8988d48c-9c55-4e0b-ad32-c37a4c0ae476.vmdk to [datastore2] c01a5d24-eb46-4a69-993e-753880ce8e85/c01a5d24-eb46-4a69-993e-753880ce8e85.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 931.283931] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0adb8970-8fa6-4b14-beb8-0fd9ded1894f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.306918] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] c01a5d24-eb46-4a69-993e-753880ce8e85/c01a5d24-eb46-4a69-993e-753880ce8e85.vmdk or device None with type streamOptimized {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.307471] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "refresh_cache-13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.309498] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d65d4338-e109-4c21-95c0-9af54d3aa711 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.330689] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 931.330689] env[63372]: value = "task-1024460" [ 931.330689] env[63372]: _type = "Task" [ 931.330689] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.340054] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024460, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.478523] env[63372]: DEBUG nova.network.neutron [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Successfully created port: 57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 931.595587] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024459, 'name': PowerOffVM_Task, 'duration_secs': 0.125022} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.595849] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 931.596077] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 931.596828] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc5a26d-ad45-4382-bdc2-3a874c2469ba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.602996] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 931.603225] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1ffd325a-58d3-4638-900b-edd00d85391f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.625689] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 931.625905] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 931.626102] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Deleting the datastore file [datastore2] 662d9fdf-ec0d-4466-8d15-abc942e4dad8 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 931.626345] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-547fd047-45a6-4a55-bf50-0d68b0449299 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.631879] env[63372]: DEBUG nova.compute.manager [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 931.635540] env[63372]: DEBUG oslo_concurrency.lockutils [None req-343f295d-fbd7-4834-b766-0d4ade757b6e tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.636618] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 931.636618] env[63372]: value = "task-1024462" [ 931.636618] env[63372]: _type = "Task" [ 931.636618] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.637025] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.659s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.638337] env[63372]: INFO nova.compute.claims [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.648359] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024462, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.683426] env[63372]: DEBUG oslo_concurrency.lockutils [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] Acquiring lock "13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.728867] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52750143-00a7-33ad-da65-93fc9a461593, 'name': SearchDatastore_Task, 'duration_secs': 0.007527} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.729686] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b250d3d-0cbd-41d4-9714-6c3bfc73e863 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.734963] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 931.734963] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b89907-912c-0587-1863-27ae22b2272c" [ 931.734963] env[63372]: _type = "Task" [ 931.734963] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.742178] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b89907-912c-0587-1863-27ae22b2272c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.825867] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1684875-dc91-409c-a43e-21f84c49c870 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-13bde897-8446-42a2-b02d-2f5b48e6f432-a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.792s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.840798] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024460, 'name': ReconfigVM_Task, 'duration_secs': 0.266259} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.841680] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Reconfigured VM instance instance-00000041 to attach disk [datastore2] c01a5d24-eb46-4a69-993e-753880ce8e85/c01a5d24-eb46-4a69-993e-753880ce8e85.vmdk or device None with type streamOptimized {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.843072] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encrypted': False, 'encryption_options': None, 'encryption_format': None, 'size': 0, 'boot_index': 0, 'disk_bus': None, 'device_type': 'disk', 'encryption_secret_uuid': None, 'device_name': '/dev/sda', 'guest_format': None, 'image_id': 'd7ae1717-77b2-47f6-9acd-b27bac221f7c'}], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'b2bc4986-aba4-402c-8024-89c05a6c57e7', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227396', 'volume_id': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'name': 'volume-cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c01a5d24-eb46-4a69-993e-753880ce8e85', 'attached_at': '', 'detached_at': '', 'volume_id': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'serial': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3'}, 'boot_index': None, 'disk_bus': None, 'device_type': None, 'mount_device': '/dev/sdb', 'guest_format': None, 'delete_on_termination': False, 'volume_type': None}], 'swap': None} {{(pid=63372) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 931.843300] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Volume attach. Driver type: vmdk {{(pid=63372) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 931.843496] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227396', 'volume_id': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'name': 'volume-cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c01a5d24-eb46-4a69-993e-753880ce8e85', 'attached_at': '', 'detached_at': '', 'volume_id': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'serial': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 931.844529] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb37dfe5-0237-4c68-b792-9335f9d1e04c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.860312] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd2512f-c88f-4e93-bf05-9e2053915821 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.886375] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] volume-cccbd23a-5f23-47d3-9e07-0c6321e2cfe3/volume-cccbd23a-5f23-47d3-9e07-0c6321e2cfe3.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.886720] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-102f6a07-7da5-47b3-8dc5-522ddbc15f97 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.905056] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 931.905056] env[63372]: value = "task-1024463" [ 931.905056] env[63372]: _type = "Task" [ 931.905056] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.912992] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024463, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.152413] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024462, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.100724} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.152684] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.152869] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 932.153080] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 932.244711] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b89907-912c-0587-1863-27ae22b2272c, 'name': SearchDatastore_Task, 'duration_secs': 0.011545} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.244969] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.245239] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 95178e2e-9d71-4606-a856-c530bfbb9345/95178e2e-9d71-4606-a856-c530bfbb9345.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 932.245483] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d454abd4-8ce6-4ced-bcfc-184af3c1bc9e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.251903] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 932.251903] env[63372]: value = "task-1024464" [ 932.251903] env[63372]: _type = "Task" [ 932.251903] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.259075] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024464, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.415110] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024463, 'name': ReconfigVM_Task, 'duration_secs': 0.285849} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.415534] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Reconfigured VM instance instance-00000041 to attach disk [datastore1] volume-cccbd23a-5f23-47d3-9e07-0c6321e2cfe3/volume-cccbd23a-5f23-47d3-9e07-0c6321e2cfe3.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.420374] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8891bf05-9c01-4539-838a-451be09b7fe5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.437162] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 932.437162] env[63372]: value = "task-1024465" [ 932.437162] env[63372]: _type = "Task" [ 932.437162] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.447373] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024465, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.646476] env[63372]: DEBUG nova.compute.manager [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 932.676158] env[63372]: DEBUG nova.virt.hardware [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 932.676443] env[63372]: DEBUG nova.virt.hardware [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 932.676609] env[63372]: DEBUG nova.virt.hardware [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 932.676791] env[63372]: DEBUG nova.virt.hardware [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 932.676992] env[63372]: DEBUG nova.virt.hardware [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 932.677096] env[63372]: DEBUG nova.virt.hardware [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 932.677396] env[63372]: DEBUG nova.virt.hardware [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 932.677625] env[63372]: DEBUG nova.virt.hardware [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 932.677817] env[63372]: DEBUG nova.virt.hardware [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 932.678320] env[63372]: DEBUG nova.virt.hardware [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 932.678551] env[63372]: DEBUG nova.virt.hardware [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 932.679475] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae36fb1-dad5-4095-aae9-3f95738e5536 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.688295] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f7e856-2837-4706-8727-e61c8eeb89a6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.731386] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 932.731614] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 932.731882] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Deleting the datastore file [datastore1] 13bde897-8446-42a2-b02d-2f5b48e6f432 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 932.732234] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cecab978-f01c-4f76-b494-cae012cca387 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.741355] env[63372]: DEBUG oslo_vmware.api [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 932.741355] env[63372]: value = "task-1024466" [ 932.741355] env[63372]: _type = "Task" [ 932.741355] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.762644] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024464, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438639} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.765481] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 95178e2e-9d71-4606-a856-c530bfbb9345/95178e2e-9d71-4606-a856-c530bfbb9345.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 932.765763] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 932.766250] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-211ff916-c863-4c19-8675-c2870bf7f9a2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.772948] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 932.772948] env[63372]: value = "task-1024467" [ 932.772948] env[63372]: _type = "Task" [ 932.772948] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.783103] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024467, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.864320] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f6b62ee-57f5-4a74-95bd-ae590da16c6d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.872249] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0be8356-d69b-4680-98c8-cb625acacaa7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.903690] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36dc337-5a13-481a-a4b0-d7a9cfde0942 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.911738] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e33401ea-da4e-4dea-9d51-b01091c0849e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.925510] env[63372]: DEBUG nova.compute.provider_tree [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.948632] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024465, 'name': ReconfigVM_Task, 'duration_secs': 0.34641} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.949174] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227396', 'volume_id': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'name': 'volume-cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c01a5d24-eb46-4a69-993e-753880ce8e85', 'attached_at': '', 'detached_at': '', 'volume_id': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'serial': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 932.949871] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d7dae72c-8ca6-4544-9879-785cdfee3396 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.958023] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 932.958023] env[63372]: value = "task-1024468" [ 932.958023] env[63372]: _type = "Task" [ 932.958023] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.966231] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024468, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.037465] env[63372]: DEBUG nova.network.neutron [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Successfully updated port: 57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 933.086188] env[63372]: DEBUG nova.compute.manager [req-02527530-a98e-432b-b86f-a6f2512f8eba req-0af90098-336a-4427-88ad-54d55633f202 service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Received event network-vif-plugged-57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.086373] env[63372]: DEBUG oslo_concurrency.lockutils [req-02527530-a98e-432b-b86f-a6f2512f8eba req-0af90098-336a-4427-88ad-54d55633f202 service nova] Acquiring lock "91985614-b959-401e-bb06-d67b230ee026-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.087126] env[63372]: DEBUG oslo_concurrency.lockutils [req-02527530-a98e-432b-b86f-a6f2512f8eba req-0af90098-336a-4427-88ad-54d55633f202 service nova] Lock "91985614-b959-401e-bb06-d67b230ee026-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.087329] env[63372]: DEBUG oslo_concurrency.lockutils [req-02527530-a98e-432b-b86f-a6f2512f8eba req-0af90098-336a-4427-88ad-54d55633f202 service nova] Lock "91985614-b959-401e-bb06-d67b230ee026-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.087554] env[63372]: DEBUG nova.compute.manager [req-02527530-a98e-432b-b86f-a6f2512f8eba req-0af90098-336a-4427-88ad-54d55633f202 service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] No waiting events found dispatching network-vif-plugged-57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 933.087764] env[63372]: WARNING nova.compute.manager [req-02527530-a98e-432b-b86f-a6f2512f8eba req-0af90098-336a-4427-88ad-54d55633f202 service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Received unexpected event network-vif-plugged-57943c40-a8c1-4d1a-bb83-d257029ce77f for instance with vm_state building and task_state spawning. [ 933.188098] env[63372]: DEBUG nova.virt.hardware [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.188375] env[63372]: DEBUG nova.virt.hardware [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.188534] env[63372]: DEBUG nova.virt.hardware [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.188722] env[63372]: DEBUG nova.virt.hardware [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.188870] env[63372]: DEBUG nova.virt.hardware [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.189022] env[63372]: DEBUG nova.virt.hardware [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.189229] env[63372]: DEBUG nova.virt.hardware [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.189393] env[63372]: DEBUG nova.virt.hardware [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.189638] env[63372]: DEBUG nova.virt.hardware [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.189804] env[63372]: DEBUG nova.virt.hardware [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.189975] env[63372]: DEBUG nova.virt.hardware [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.190883] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a987f6d0-5be8-4865-a7c7-165d32642e4d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.198705] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7039d8fe-db86-456a-9679-c70a4a85ceb2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.211976] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Instance VIF info [] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 933.217499] env[63372]: DEBUG oslo.service.loopingcall [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.218051] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 933.218266] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-628753dd-c9cb-46b1-99eb-d1ae6456427b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.234951] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 933.234951] env[63372]: value = "task-1024469" [ 933.234951] env[63372]: _type = "Task" [ 933.234951] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.242569] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024469, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.249983] env[63372]: DEBUG oslo_vmware.api [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024466, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217378} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.250270] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 933.250491] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 933.250704] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 933.250885] env[63372]: INFO nova.compute.manager [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Took 3.52 seconds to destroy the instance on the hypervisor. [ 933.251211] env[63372]: DEBUG oslo.service.loopingcall [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.251327] env[63372]: DEBUG nova.compute.manager [-] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 933.251419] env[63372]: DEBUG nova.network.neutron [-] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 933.281391] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024467, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065267} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.283699] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 933.284501] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab86bf1f-ab27-4d8e-a714-7a97da50e144 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.306910] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 95178e2e-9d71-4606-a856-c530bfbb9345/95178e2e-9d71-4606-a856-c530bfbb9345.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.307673] env[63372]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port eac444f7-19cf-4cc8-b354-21b08d90753c could not be found.", "detail": ""}} {{(pid=63372) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 933.307891] env[63372]: DEBUG nova.network.neutron [-] Unable to show port eac444f7-19cf-4cc8-b354-21b08d90753c as it no longer exists. {{(pid=63372) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 933.309284] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8636767e-e0df-46a4-a06f-6cd982193336 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.329287] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 933.329287] env[63372]: value = "task-1024470" [ 933.329287] env[63372]: _type = "Task" [ 933.329287] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.337889] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024470, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.428703] env[63372]: DEBUG nova.scheduler.client.report [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.466651] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024468, 'name': Rename_Task, 'duration_secs': 0.163217} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.466953] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 933.467261] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-975ba1f6-ac28-4d84-969b-bd413c165e74 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.473364] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 933.473364] env[63372]: value = "task-1024471" [ 933.473364] env[63372]: _type = "Task" [ 933.473364] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.484705] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024471, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.540215] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.540415] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquired lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.540938] env[63372]: DEBUG nova.network.neutron [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 933.745245] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024469, 'name': CreateVM_Task, 'duration_secs': 0.304005} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.745431] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 933.745886] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.746071] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.746452] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 933.746642] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a413844-ce9c-4553-8405-63fce3f1f409 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.750977] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 933.750977] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52ea94b0-c1f1-28ab-8d7f-dc056b16f54b" [ 933.750977] env[63372]: _type = "Task" [ 933.750977] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.759860] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ea94b0-c1f1-28ab-8d7f-dc056b16f54b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.838346] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024470, 'name': ReconfigVM_Task, 'duration_secs': 0.382646} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.838689] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 95178e2e-9d71-4606-a856-c530bfbb9345/95178e2e-9d71-4606-a856-c530bfbb9345.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.839271] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-66eeaf96-dc18-4f14-972e-505e5f0a15a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.846423] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 933.846423] env[63372]: value = "task-1024472" [ 933.846423] env[63372]: _type = "Task" [ 933.846423] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.854215] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024472, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.934065] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.934645] env[63372]: DEBUG nova.compute.manager [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 933.938585] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.479s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.938841] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.941227] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.422s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.943015] env[63372]: INFO nova.compute.claims [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 933.967462] env[63372]: INFO nova.scheduler.client.report [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleted allocations for instance d568b727-7836-4b1d-9f27-f159227e46f7 [ 933.984050] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024471, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.071784] env[63372]: DEBUG nova.network.neutron [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 934.263244] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ea94b0-c1f1-28ab-8d7f-dc056b16f54b, 'name': SearchDatastore_Task, 'duration_secs': 0.008238} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.263426] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.263665] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 934.263903] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.264068] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.264528] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 934.264528] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-046aa287-b2cb-41f4-a484-6b58124648a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.273827] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 934.274015] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 934.274797] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2acd0c51-aafd-4221-a7b6-8ed3b3fbd572 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.280491] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 934.280491] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5241033c-884f-575b-1440-1bf7b63cc6f8" [ 934.280491] env[63372]: _type = "Task" [ 934.280491] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.288169] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5241033c-884f-575b-1440-1bf7b63cc6f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.297698] env[63372]: DEBUG nova.network.neutron [-] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.344549] env[63372]: DEBUG nova.network.neutron [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updating instance_info_cache with network_info: [{"id": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "address": "fa:16:3e:1f:74:b1", "network": {"id": "cbdcb945-f0a5-49b1-9b18-fcd8e4ca9728", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1068757189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e517915d5e7a4e3e924cc95c55170a66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57943c40-a8", "ovs_interfaceid": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.357492] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024472, 'name': Rename_Task, 'duration_secs': 0.168426} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.357492] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 934.357492] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c317b29-7e00-47b2-9ad4-4da49e742015 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.364791] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 934.364791] env[63372]: value = "task-1024473" [ 934.364791] env[63372]: _type = "Task" [ 934.364791] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.376753] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024473, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.440493] env[63372]: DEBUG nova.compute.utils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.441930] env[63372]: DEBUG nova.compute.manager [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 934.442112] env[63372]: DEBUG nova.network.neutron [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 934.478523] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3ac60e9-d286-40db-a5a4-bc8a370495ea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "d568b727-7836-4b1d-9f27-f159227e46f7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.189s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.485429] env[63372]: DEBUG oslo_vmware.api [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024471, 'name': PowerOnVM_Task, 'duration_secs': 0.622636} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.487073] env[63372]: DEBUG nova.policy [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'be1b6f7a5f954f41b855e97d19b0b214', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8a468e54b8b4ec28f3fd684c7695ce1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 934.488663] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 934.592458] env[63372]: DEBUG nova.compute.manager [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 934.593122] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d672b1-0cf8-49cb-8690-de9daa5eb314 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.793222] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5241033c-884f-575b-1440-1bf7b63cc6f8, 'name': SearchDatastore_Task, 'duration_secs': 0.03429} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.793890] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0d61fe4-a93f-468d-8bda-f462d16dd44c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.800471] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 934.800471] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e0f8ef-011a-fd55-30fa-d58aa70f5c00" [ 934.800471] env[63372]: _type = "Task" [ 934.800471] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.802875] env[63372]: INFO nova.compute.manager [-] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Took 1.55 seconds to deallocate network for instance. [ 934.814049] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e0f8ef-011a-fd55-30fa-d58aa70f5c00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.850023] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Releasing lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.850023] env[63372]: DEBUG nova.compute.manager [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Instance network_info: |[{"id": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "address": "fa:16:3e:1f:74:b1", "network": {"id": "cbdcb945-f0a5-49b1-9b18-fcd8e4ca9728", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1068757189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e517915d5e7a4e3e924cc95c55170a66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57943c40-a8", "ovs_interfaceid": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 934.850023] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:74:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd3c6b64-aba2-4bdc-a693-3b4dff3ed861', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '57943c40-a8c1-4d1a-bb83-d257029ce77f', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 934.859101] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Creating folder: Project (e517915d5e7a4e3e924cc95c55170a66). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 934.859101] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2fca0205-c839-4ddb-8f04-11f7f35e2f0f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.874134] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Created folder: Project (e517915d5e7a4e3e924cc95c55170a66) in parent group-v227230. [ 934.876025] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Creating folder: Instances. Parent ref: group-v227415. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 934.876025] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9fe32a82-42e5-444f-8575-f09a254e4fa2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.880627] env[63372]: DEBUG oslo_vmware.api [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024473, 'name': PowerOnVM_Task, 'duration_secs': 0.474124} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.881039] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 934.881361] env[63372]: INFO nova.compute.manager [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Took 8.88 seconds to spawn the instance on the hypervisor. [ 934.881741] env[63372]: DEBUG nova.compute.manager [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 934.882589] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f29f0a-1f40-446a-a35a-df858d139fcd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.888017] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Created folder: Instances in parent group-v227415. [ 934.888017] env[63372]: DEBUG oslo.service.loopingcall [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.888017] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91985614-b959-401e-bb06-d67b230ee026] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 934.892538] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1aea62f3-b9b4-45e8-a49e-043f0c531700 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.910888] env[63372]: DEBUG nova.network.neutron [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Successfully created port: e129d65d-67e7-49fb-95d6-2b57141bd3b3 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.918895] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 934.918895] env[63372]: value = "task-1024476" [ 934.918895] env[63372]: _type = "Task" [ 934.918895] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.926640] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024476, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.946195] env[63372]: DEBUG nova.compute.manager [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 935.114900] env[63372]: DEBUG oslo_concurrency.lockutils [None req-25e0bb82-f4d9-4e3a-a593-d354788fe5d1 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 46.610s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.120937] env[63372]: DEBUG nova.compute.manager [req-c9db16e9-1136-4742-8924-a507b4f07303 req-6af13fe8-ae50-49fd-bcd3-6543f1beb804 service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Received event network-changed-57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.121248] env[63372]: DEBUG nova.compute.manager [req-c9db16e9-1136-4742-8924-a507b4f07303 req-6af13fe8-ae50-49fd-bcd3-6543f1beb804 service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Refreshing instance network info cache due to event network-changed-57943c40-a8c1-4d1a-bb83-d257029ce77f. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 935.121355] env[63372]: DEBUG oslo_concurrency.lockutils [req-c9db16e9-1136-4742-8924-a507b4f07303 req-6af13fe8-ae50-49fd-bcd3-6543f1beb804 service nova] Acquiring lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.121499] env[63372]: DEBUG oslo_concurrency.lockutils [req-c9db16e9-1136-4742-8924-a507b4f07303 req-6af13fe8-ae50-49fd-bcd3-6543f1beb804 service nova] Acquired lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.121664] env[63372]: DEBUG nova.network.neutron [req-c9db16e9-1136-4742-8924-a507b4f07303 req-6af13fe8-ae50-49fd-bcd3-6543f1beb804 service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Refreshing network info cache for port 57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 935.169431] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-497860f8-bec4-469c-b288-3b8576129327 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.177420] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b79d2a58-dac4-4d71-96e2-b1e331b5f301 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.210091] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e377a1-5a0a-496d-8162-542520355fb2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.218590] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d1f03a4-26a8-4f76-942c-0bd06b2a0250 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.232964] env[63372]: DEBUG nova.compute.provider_tree [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 935.313635] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e0f8ef-011a-fd55-30fa-d58aa70f5c00, 'name': SearchDatastore_Task, 'duration_secs': 0.022737} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.313963] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.314257] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 662d9fdf-ec0d-4466-8d15-abc942e4dad8/662d9fdf-ec0d-4466-8d15-abc942e4dad8.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 935.315285] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.315492] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b1227454-0c98-4b90-9f41-f014052f5755 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.323470] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 935.323470] env[63372]: value = "task-1024477" [ 935.323470] env[63372]: _type = "Task" [ 935.323470] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.332225] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024477, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.424252] env[63372]: INFO nova.compute.manager [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Took 26.40 seconds to build instance. [ 935.431191] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024476, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.759097] env[63372]: ERROR nova.scheduler.client.report [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [req-e2664b2d-7bad-4683-8b58-60fd86a5ec5f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e2664b2d-7bad-4683-8b58-60fd86a5ec5f"}]} [ 935.785061] env[63372]: DEBUG nova.scheduler.client.report [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Refreshing inventories for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 935.806871] env[63372]: DEBUG nova.scheduler.client.report [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Updating ProviderTree inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 935.807168] env[63372]: DEBUG nova.compute.provider_tree [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 935.826654] env[63372]: DEBUG nova.scheduler.client.report [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Refreshing aggregate associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, aggregates: None {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 935.844557] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024477, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.863255] env[63372]: DEBUG nova.scheduler.client.report [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Refreshing trait associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 935.926719] env[63372]: DEBUG oslo_concurrency.lockutils [None req-73a44cc6-d5a5-43b1-a71b-d8b66b85d115 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "95178e2e-9d71-4606-a856-c530bfbb9345" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.910s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.934476] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024476, 'name': CreateVM_Task, 'duration_secs': 0.554139} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.934648] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91985614-b959-401e-bb06-d67b230ee026] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 935.935814] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.935996] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.936332] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 935.936636] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2de1ed9-9dac-424b-bc33-7129c3a0b6c5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.945793] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 935.945793] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523e5ea4-e102-d5ac-a6da-dc920dc11a8b" [ 935.945793] env[63372]: _type = "Task" [ 935.945793] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.955032] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523e5ea4-e102-d5ac-a6da-dc920dc11a8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.959905] env[63372]: DEBUG nova.compute.manager [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 935.989015] env[63372]: DEBUG nova.virt.hardware [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.989280] env[63372]: DEBUG nova.virt.hardware [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.989440] env[63372]: DEBUG nova.virt.hardware [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.989630] env[63372]: DEBUG nova.virt.hardware [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.989778] env[63372]: DEBUG nova.virt.hardware [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.989927] env[63372]: DEBUG nova.virt.hardware [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.990284] env[63372]: DEBUG nova.virt.hardware [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.990475] env[63372]: DEBUG nova.virt.hardware [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.990715] env[63372]: DEBUG nova.virt.hardware [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.991052] env[63372]: DEBUG nova.virt.hardware [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.991267] env[63372]: DEBUG nova.virt.hardware [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.992259] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b03690c-e716-42e9-8d47-0ea06a04a287 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.005429] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776fa854-b3a8-4744-8500-576c25230870 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.073841] env[63372]: DEBUG nova.network.neutron [req-c9db16e9-1136-4742-8924-a507b4f07303 req-6af13fe8-ae50-49fd-bcd3-6543f1beb804 service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updated VIF entry in instance network info cache for port 57943c40-a8c1-4d1a-bb83-d257029ce77f. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 936.074032] env[63372]: DEBUG nova.network.neutron [req-c9db16e9-1136-4742-8924-a507b4f07303 req-6af13fe8-ae50-49fd-bcd3-6543f1beb804 service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updating instance_info_cache with network_info: [{"id": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "address": "fa:16:3e:1f:74:b1", "network": {"id": "cbdcb945-f0a5-49b1-9b18-fcd8e4ca9728", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1068757189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e517915d5e7a4e3e924cc95c55170a66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57943c40-a8", "ovs_interfaceid": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.115560] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d349412-ebab-4e37-9dba-d3fa5ceaa099 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.124149] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914a1037-bced-4a89-bbe1-71afed1f4927 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.158781] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac0d196-e1d2-4071-9b6b-d151016e164a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.167449] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f97756aa-061d-4916-9092-e8b800e64e75 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.183167] env[63372]: DEBUG nova.compute.provider_tree [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.337171] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024477, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.64209} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.337494] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 662d9fdf-ec0d-4466-8d15-abc942e4dad8/662d9fdf-ec0d-4466-8d15-abc942e4dad8.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 936.337757] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.338064] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ec899655-e286-4a4a-b938-a7a9b7cb6e17 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.345244] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 936.345244] env[63372]: value = "task-1024478" [ 936.345244] env[63372]: _type = "Task" [ 936.345244] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.354225] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024478, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.456052] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523e5ea4-e102-d5ac-a6da-dc920dc11a8b, 'name': SearchDatastore_Task, 'duration_secs': 0.015548} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.456385] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.456624] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 936.456865] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.457033] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.457213] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.457464] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-efc04373-b626-4793-9a61-81460a0fb5fa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.467195] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.467382] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 936.468190] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64914171-375a-4ff2-abb3-1dd277ab974f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.473231] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 936.473231] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fe3f54-2616-0c1a-8754-17d70229f833" [ 936.473231] env[63372]: _type = "Task" [ 936.473231] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.480960] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fe3f54-2616-0c1a-8754-17d70229f833, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.538834] env[63372]: DEBUG nova.compute.manager [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Stashing vm_state: active {{(pid=63372) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 936.579764] env[63372]: DEBUG oslo_concurrency.lockutils [req-c9db16e9-1136-4742-8924-a507b4f07303 req-6af13fe8-ae50-49fd-bcd3-6543f1beb804 service nova] Releasing lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.580093] env[63372]: DEBUG nova.compute.manager [req-c9db16e9-1136-4742-8924-a507b4f07303 req-6af13fe8-ae50-49fd-bcd3-6543f1beb804 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Received event network-vif-deleted-b02b2fb0-9262-4f24-a677-c4f61362410a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.616762] env[63372]: DEBUG nova.network.neutron [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Successfully updated port: e129d65d-67e7-49fb-95d6-2b57141bd3b3 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 936.635242] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "95178e2e-9d71-4606-a856-c530bfbb9345" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.635242] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "95178e2e-9d71-4606-a856-c530bfbb9345" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.635242] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "95178e2e-9d71-4606-a856-c530bfbb9345-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.635242] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "95178e2e-9d71-4606-a856-c530bfbb9345-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.635242] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "95178e2e-9d71-4606-a856-c530bfbb9345-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.636360] env[63372]: INFO nova.compute.manager [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Terminating instance [ 936.638788] env[63372]: DEBUG nova.compute.manager [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 936.638982] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 936.640394] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb1f0859-45f7-43dd-9199-73e6c8e4fd65 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.648342] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 936.648607] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc1dd4c0-4efd-4ffa-8477-2327aff641a6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.654974] env[63372]: DEBUG oslo_vmware.api [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 936.654974] env[63372]: value = "task-1024479" [ 936.654974] env[63372]: _type = "Task" [ 936.654974] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.663319] env[63372]: DEBUG oslo_vmware.api [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024479, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.686396] env[63372]: DEBUG nova.scheduler.client.report [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 936.855799] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024478, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.251509} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.856164] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 936.856970] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e977da-dcdf-4179-9b12-8f3c58275cfa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.877979] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 662d9fdf-ec0d-4466-8d15-abc942e4dad8/662d9fdf-ec0d-4466-8d15-abc942e4dad8.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.878357] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fcc11405-9fcb-4b88-a8d1-d304e0ec4252 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.898398] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 936.898398] env[63372]: value = "task-1024480" [ 936.898398] env[63372]: _type = "Task" [ 936.898398] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.908479] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024480, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.986057] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fe3f54-2616-0c1a-8754-17d70229f833, 'name': SearchDatastore_Task, 'duration_secs': 0.012544} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.987026] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2343e7c3-7137-434d-b0bf-0fc2828a8ef2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.992724] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 936.992724] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]526b198d-99a1-1b4e-0315-cd503f8a3bf0" [ 936.992724] env[63372]: _type = "Task" [ 936.992724] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.000771] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526b198d-99a1-1b4e-0315-cd503f8a3bf0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.060079] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.121564] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Acquiring lock "refresh_cache-ed2cefa1-3604-43c5-9860-0fb49133bbb4" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.121716] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Acquired lock "refresh_cache-ed2cefa1-3604-43c5-9860-0fb49133bbb4" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.121945] env[63372]: DEBUG nova.network.neutron [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 937.165410] env[63372]: DEBUG oslo_vmware.api [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024479, 'name': PowerOffVM_Task, 'duration_secs': 0.226727} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.165703] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 937.165874] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 937.166136] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1bac7902-8097-4c1b-9e91-4ece7522c14c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.192287] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.251s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.192788] env[63372]: DEBUG nova.compute.manager [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 937.195468] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 12.550s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.195662] env[63372]: DEBUG nova.objects.instance [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63372) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 937.214979] env[63372]: DEBUG nova.compute.manager [req-2d1ff9dc-6c0a-4e82-862d-c78fc0d7e48d req-b0b39761-c080-4e6f-b6ec-9b3705cfe63d service nova] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Received event network-vif-plugged-e129d65d-67e7-49fb-95d6-2b57141bd3b3 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.215586] env[63372]: DEBUG oslo_concurrency.lockutils [req-2d1ff9dc-6c0a-4e82-862d-c78fc0d7e48d req-b0b39761-c080-4e6f-b6ec-9b3705cfe63d service nova] Acquiring lock "ed2cefa1-3604-43c5-9860-0fb49133bbb4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.215814] env[63372]: DEBUG oslo_concurrency.lockutils [req-2d1ff9dc-6c0a-4e82-862d-c78fc0d7e48d req-b0b39761-c080-4e6f-b6ec-9b3705cfe63d service nova] Lock "ed2cefa1-3604-43c5-9860-0fb49133bbb4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.216016] env[63372]: DEBUG oslo_concurrency.lockutils [req-2d1ff9dc-6c0a-4e82-862d-c78fc0d7e48d req-b0b39761-c080-4e6f-b6ec-9b3705cfe63d service nova] Lock "ed2cefa1-3604-43c5-9860-0fb49133bbb4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.216378] env[63372]: DEBUG nova.compute.manager [req-2d1ff9dc-6c0a-4e82-862d-c78fc0d7e48d req-b0b39761-c080-4e6f-b6ec-9b3705cfe63d service nova] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] No waiting events found dispatching network-vif-plugged-e129d65d-67e7-49fb-95d6-2b57141bd3b3 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 937.216810] env[63372]: WARNING nova.compute.manager [req-2d1ff9dc-6c0a-4e82-862d-c78fc0d7e48d req-b0b39761-c080-4e6f-b6ec-9b3705cfe63d service nova] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Received unexpected event network-vif-plugged-e129d65d-67e7-49fb-95d6-2b57141bd3b3 for instance with vm_state building and task_state spawning. [ 937.216810] env[63372]: DEBUG nova.compute.manager [req-2d1ff9dc-6c0a-4e82-862d-c78fc0d7e48d req-b0b39761-c080-4e6f-b6ec-9b3705cfe63d service nova] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Received event network-changed-e129d65d-67e7-49fb-95d6-2b57141bd3b3 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.216935] env[63372]: DEBUG nova.compute.manager [req-2d1ff9dc-6c0a-4e82-862d-c78fc0d7e48d req-b0b39761-c080-4e6f-b6ec-9b3705cfe63d service nova] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Refreshing instance network info cache due to event network-changed-e129d65d-67e7-49fb-95d6-2b57141bd3b3. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.217051] env[63372]: DEBUG oslo_concurrency.lockutils [req-2d1ff9dc-6c0a-4e82-862d-c78fc0d7e48d req-b0b39761-c080-4e6f-b6ec-9b3705cfe63d service nova] Acquiring lock "refresh_cache-ed2cefa1-3604-43c5-9860-0fb49133bbb4" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.239027] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 937.239179] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 937.239385] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleting the datastore file [datastore2] 95178e2e-9d71-4606-a856-c530bfbb9345 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 937.239888] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98316d70-ed0f-440f-9046-3e063a73243d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.247418] env[63372]: DEBUG oslo_vmware.api [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 937.247418] env[63372]: value = "task-1024482" [ 937.247418] env[63372]: _type = "Task" [ 937.247418] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.258959] env[63372]: DEBUG oslo_vmware.api [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024482, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.408208] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024480, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.503335] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526b198d-99a1-1b4e-0315-cd503f8a3bf0, 'name': SearchDatastore_Task, 'duration_secs': 0.044779} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.503955] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.503955] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 91985614-b959-401e-bb06-d67b230ee026/91985614-b959-401e-bb06-d67b230ee026.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 937.504326] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5721a025-44e5-4cd8-ad08-59eade387714 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.511771] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 937.511771] env[63372]: value = "task-1024483" [ 937.511771] env[63372]: _type = "Task" [ 937.511771] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.519569] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024483, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.652665] env[63372]: DEBUG nova.network.neutron [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 937.703798] env[63372]: DEBUG nova.compute.utils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 937.705662] env[63372]: DEBUG nova.compute.manager [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 937.705927] env[63372]: DEBUG nova.network.neutron [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 937.757616] env[63372]: DEBUG oslo_vmware.api [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024482, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.486169} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.757972] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 937.758086] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 937.758270] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 937.758448] env[63372]: INFO nova.compute.manager [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Took 1.12 seconds to destroy the instance on the hypervisor. [ 937.758702] env[63372]: DEBUG oslo.service.loopingcall [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.758888] env[63372]: DEBUG nova.compute.manager [-] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 937.758983] env[63372]: DEBUG nova.network.neutron [-] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 937.811536] env[63372]: DEBUG nova.policy [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '118d2f1de2be45cc8bb48bb75525d37e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'afe39cfedf214d50be775ef736f94da9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 937.911709] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024480, 'name': ReconfigVM_Task, 'duration_secs': 0.581925} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.912060] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 662d9fdf-ec0d-4466-8d15-abc942e4dad8/662d9fdf-ec0d-4466-8d15-abc942e4dad8.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.912856] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-68df87ee-28bd-43a2-9cd0-81dbcbeda2ae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.920248] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 937.920248] env[63372]: value = "task-1024484" [ 937.920248] env[63372]: _type = "Task" [ 937.920248] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.929899] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024484, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.930960] env[63372]: DEBUG nova.network.neutron [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Updating instance_info_cache with network_info: [{"id": "e129d65d-67e7-49fb-95d6-2b57141bd3b3", "address": "fa:16:3e:c4:c1:86", "network": {"id": "7e7bc2c7-0a17-407c-a4a6-8b4c83d77519", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1637587998-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8a468e54b8b4ec28f3fd684c7695ce1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape129d65d-67", "ovs_interfaceid": "e129d65d-67e7-49fb-95d6-2b57141bd3b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.023399] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024483, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.073292] env[63372]: DEBUG nova.compute.manager [req-5c6a5a58-59d6-40bd-8a6c-bb4739aeea4c req-d00adc08-ea45-4a33-9712-07cf71d8e29d service nova] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Received event network-vif-deleted-974ab817-fe7c-424a-9174-137c3ee524f4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.073487] env[63372]: INFO nova.compute.manager [req-5c6a5a58-59d6-40bd-8a6c-bb4739aeea4c req-d00adc08-ea45-4a33-9712-07cf71d8e29d service nova] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Neutron deleted interface 974ab817-fe7c-424a-9174-137c3ee524f4; detaching it from the instance and deleting it from the info cache [ 938.073659] env[63372]: DEBUG nova.network.neutron [req-5c6a5a58-59d6-40bd-8a6c-bb4739aeea4c req-d00adc08-ea45-4a33-9712-07cf71d8e29d service nova] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.141011] env[63372]: DEBUG nova.network.neutron [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Successfully created port: a6afe51b-6c5f-4f54-9657-97c43e85cdba {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 938.207022] env[63372]: DEBUG nova.compute.manager [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 938.214183] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e8ea40e3-c386-4f28-96d6-8285a5b2fbf7 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.214183] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.410s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.214183] env[63372]: DEBUG nova.objects.instance [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'resources' on Instance uuid 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.432367] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024484, 'name': Rename_Task, 'duration_secs': 0.14497} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.432367] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 938.432367] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7a38c7f5-e995-4aa2-a5fb-a4cd20567f74 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.437018] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Releasing lock "refresh_cache-ed2cefa1-3604-43c5-9860-0fb49133bbb4" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.437018] env[63372]: DEBUG nova.compute.manager [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Instance network_info: |[{"id": "e129d65d-67e7-49fb-95d6-2b57141bd3b3", "address": "fa:16:3e:c4:c1:86", "network": {"id": "7e7bc2c7-0a17-407c-a4a6-8b4c83d77519", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1637587998-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8a468e54b8b4ec28f3fd684c7695ce1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape129d65d-67", "ovs_interfaceid": "e129d65d-67e7-49fb-95d6-2b57141bd3b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 938.437018] env[63372]: DEBUG oslo_concurrency.lockutils [req-2d1ff9dc-6c0a-4e82-862d-c78fc0d7e48d req-b0b39761-c080-4e6f-b6ec-9b3705cfe63d service nova] Acquired lock "refresh_cache-ed2cefa1-3604-43c5-9860-0fb49133bbb4" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.437018] env[63372]: DEBUG nova.network.neutron [req-2d1ff9dc-6c0a-4e82-862d-c78fc0d7e48d req-b0b39761-c080-4e6f-b6ec-9b3705cfe63d service nova] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Refreshing network info cache for port e129d65d-67e7-49fb-95d6-2b57141bd3b3 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 938.437018] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:c1:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a34aa30-95be-4b18-98ca-1f2d81f7e9e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e129d65d-67e7-49fb-95d6-2b57141bd3b3', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.444188] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Creating folder: Project (d8a468e54b8b4ec28f3fd684c7695ce1). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 938.448280] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-75596f9e-01b4-4d64-bda9-708fa97de8b0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.450392] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 938.450392] env[63372]: value = "task-1024485" [ 938.450392] env[63372]: _type = "Task" [ 938.450392] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.460863] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024485, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.462219] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Created folder: Project (d8a468e54b8b4ec28f3fd684c7695ce1) in parent group-v227230. [ 938.462555] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Creating folder: Instances. Parent ref: group-v227418. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 938.462885] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1760ed4-f8b1-42a3-9ced-b5f74eb9a74d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.473266] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Created folder: Instances in parent group-v227418. [ 938.473641] env[63372]: DEBUG oslo.service.loopingcall [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.473954] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 938.474355] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d77a54a-5d47-45ad-8247-6293037b1b72 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.497749] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.497749] env[63372]: value = "task-1024488" [ 938.497749] env[63372]: _type = "Task" [ 938.497749] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.503852] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024488, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.528484] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024483, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.830874} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.531140] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 91985614-b959-401e-bb06-d67b230ee026/91985614-b959-401e-bb06-d67b230ee026.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 938.531530] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 938.532051] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38b09be0-9744-42d1-b1bd-4d983f42ac65 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.538093] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 938.538093] env[63372]: value = "task-1024489" [ 938.538093] env[63372]: _type = "Task" [ 938.538093] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.542080] env[63372]: DEBUG nova.network.neutron [-] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.547044] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024489, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.575889] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f07de979-23d4-486e-a986-53babc4372cc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.585232] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c04d24-a077-4694-bbc5-1fd229476dc0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.615669] env[63372]: DEBUG nova.compute.manager [req-5c6a5a58-59d6-40bd-8a6c-bb4739aeea4c req-d00adc08-ea45-4a33-9712-07cf71d8e29d service nova] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Detach interface failed, port_id=974ab817-fe7c-424a-9174-137c3ee524f4, reason: Instance 95178e2e-9d71-4606-a856-c530bfbb9345 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 938.679123] env[63372]: DEBUG nova.network.neutron [req-2d1ff9dc-6c0a-4e82-862d-c78fc0d7e48d req-b0b39761-c080-4e6f-b6ec-9b3705cfe63d service nova] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Updated VIF entry in instance network info cache for port e129d65d-67e7-49fb-95d6-2b57141bd3b3. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 938.680055] env[63372]: DEBUG nova.network.neutron [req-2d1ff9dc-6c0a-4e82-862d-c78fc0d7e48d req-b0b39761-c080-4e6f-b6ec-9b3705cfe63d service nova] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Updating instance_info_cache with network_info: [{"id": "e129d65d-67e7-49fb-95d6-2b57141bd3b3", "address": "fa:16:3e:c4:c1:86", "network": {"id": "7e7bc2c7-0a17-407c-a4a6-8b4c83d77519", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1637587998-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8a468e54b8b4ec28f3fd684c7695ce1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape129d65d-67", "ovs_interfaceid": "e129d65d-67e7-49fb-95d6-2b57141bd3b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.901351] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76cdfeb0-7f32-44f1-9059-e830a3aba30f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.911310] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c115c893-7a2e-47c0-88cf-977facc3906c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.951027] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691773a3-92da-47b1-a93d-79e21eaeaa23 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.964511] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d0148c-19f1-4250-b7a1-858fe0298b66 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.968399] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024485, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.981657] env[63372]: DEBUG nova.compute.provider_tree [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 939.007109] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024488, 'name': CreateVM_Task, 'duration_secs': 0.393614} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.007261] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 939.008082] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.008323] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.008683] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 939.008961] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40ab242d-72b4-4dd5-aea8-7958b4539761 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.013760] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Waiting for the task: (returnval){ [ 939.013760] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f4f151-3d30-4bf9-b9c3-d271e750b8de" [ 939.013760] env[63372]: _type = "Task" [ 939.013760] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.021732] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f4f151-3d30-4bf9-b9c3-d271e750b8de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.044885] env[63372]: INFO nova.compute.manager [-] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Took 1.29 seconds to deallocate network for instance. [ 939.049998] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024489, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072752} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.052952] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.053822] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ada1a9-e826-4a53-b946-e9eb71ca7936 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.075986] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 91985614-b959-401e-bb06-d67b230ee026/91985614-b959-401e-bb06-d67b230ee026.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.076284] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a216828c-2965-49f4-9004-8a456244ac5e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.096081] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 939.096081] env[63372]: value = "task-1024490" [ 939.096081] env[63372]: _type = "Task" [ 939.096081] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.103832] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024490, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.183201] env[63372]: DEBUG oslo_concurrency.lockutils [req-2d1ff9dc-6c0a-4e82-862d-c78fc0d7e48d req-b0b39761-c080-4e6f-b6ec-9b3705cfe63d service nova] Releasing lock "refresh_cache-ed2cefa1-3604-43c5-9860-0fb49133bbb4" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.219975] env[63372]: DEBUG nova.compute.manager [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 939.244758] env[63372]: DEBUG nova.virt.hardware [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.245014] env[63372]: DEBUG nova.virt.hardware [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.245183] env[63372]: DEBUG nova.virt.hardware [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.245373] env[63372]: DEBUG nova.virt.hardware [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.245521] env[63372]: DEBUG nova.virt.hardware [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.245669] env[63372]: DEBUG nova.virt.hardware [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.245880] env[63372]: DEBUG nova.virt.hardware [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.246073] env[63372]: DEBUG nova.virt.hardware [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.246656] env[63372]: DEBUG nova.virt.hardware [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.246656] env[63372]: DEBUG nova.virt.hardware [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.246656] env[63372]: DEBUG nova.virt.hardware [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.247459] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4ba6bd-b806-453f-b67f-5fab94eb7846 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.255743] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6660050e-488a-40a2-871c-07a91d8af34a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.463043] env[63372]: DEBUG oslo_vmware.api [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024485, 'name': PowerOnVM_Task, 'duration_secs': 1.018863} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.463449] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 939.463739] env[63372]: DEBUG nova.compute.manager [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.464529] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593b08e9-cb68-4fd4-ae3b-3292a505536c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.511998] env[63372]: DEBUG nova.scheduler.client.report [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 119 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 939.512298] env[63372]: DEBUG nova.compute.provider_tree [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 119 to 120 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 939.512483] env[63372]: DEBUG nova.compute.provider_tree [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 939.525406] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f4f151-3d30-4bf9-b9c3-d271e750b8de, 'name': SearchDatastore_Task, 'duration_secs': 0.033408} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.525657] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.525889] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.526142] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.526295] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.526473] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.527233] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-772f7cd0-4389-42de-99bf-83193fb6e5ed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.538244] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.538430] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 939.539152] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ec1cb42-7c3c-47b0-873b-7a58865c3b46 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.546997] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Waiting for the task: (returnval){ [ 939.546997] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523cf130-7ee3-2fd6-d44c-52ca140bdc17" [ 939.546997] env[63372]: _type = "Task" [ 939.546997] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.555796] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523cf130-7ee3-2fd6-d44c-52ca140bdc17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.558336] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.605712] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024490, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.744864] env[63372]: DEBUG nova.compute.manager [req-cc7b4a29-14c8-4f86-9eb8-0e3104b2f012 req-cd0faab5-4488-4faf-b81f-53b80bd28965 service nova] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Received event network-vif-plugged-a6afe51b-6c5f-4f54-9657-97c43e85cdba {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.745423] env[63372]: DEBUG oslo_concurrency.lockutils [req-cc7b4a29-14c8-4f86-9eb8-0e3104b2f012 req-cd0faab5-4488-4faf-b81f-53b80bd28965 service nova] Acquiring lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.745423] env[63372]: DEBUG oslo_concurrency.lockutils [req-cc7b4a29-14c8-4f86-9eb8-0e3104b2f012 req-cd0faab5-4488-4faf-b81f-53b80bd28965 service nova] Lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.745540] env[63372]: DEBUG oslo_concurrency.lockutils [req-cc7b4a29-14c8-4f86-9eb8-0e3104b2f012 req-cd0faab5-4488-4faf-b81f-53b80bd28965 service nova] Lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.745642] env[63372]: DEBUG nova.compute.manager [req-cc7b4a29-14c8-4f86-9eb8-0e3104b2f012 req-cd0faab5-4488-4faf-b81f-53b80bd28965 service nova] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] No waiting events found dispatching network-vif-plugged-a6afe51b-6c5f-4f54-9657-97c43e85cdba {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 939.745809] env[63372]: WARNING nova.compute.manager [req-cc7b4a29-14c8-4f86-9eb8-0e3104b2f012 req-cd0faab5-4488-4faf-b81f-53b80bd28965 service nova] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Received unexpected event network-vif-plugged-a6afe51b-6c5f-4f54-9657-97c43e85cdba for instance with vm_state building and task_state spawning. [ 939.915327] env[63372]: DEBUG nova.network.neutron [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Successfully updated port: a6afe51b-6c5f-4f54-9657-97c43e85cdba {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 939.983512] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.020927] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.807s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.023312] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.208s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.024835] env[63372]: INFO nova.compute.claims [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 940.041266] env[63372]: INFO nova.scheduler.client.report [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleted allocations for instance 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5 [ 940.056593] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523cf130-7ee3-2fd6-d44c-52ca140bdc17, 'name': SearchDatastore_Task, 'duration_secs': 0.058151} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.057397] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c142b9f0-5f76-430a-a39b-bccea8ee8bd1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.063396] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Waiting for the task: (returnval){ [ 940.063396] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e2f6c7-ba69-228c-1049-37c4964b5069" [ 940.063396] env[63372]: _type = "Task" [ 940.063396] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.071731] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e2f6c7-ba69-228c-1049-37c4964b5069, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.106079] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024490, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.417614] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "refresh_cache-4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.417799] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "refresh_cache-4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.417918] env[63372]: DEBUG nova.network.neutron [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 940.430604] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquiring lock "662d9fdf-ec0d-4466-8d15-abc942e4dad8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.430847] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Lock "662d9fdf-ec0d-4466-8d15-abc942e4dad8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.431073] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquiring lock "662d9fdf-ec0d-4466-8d15-abc942e4dad8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.431266] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Lock "662d9fdf-ec0d-4466-8d15-abc942e4dad8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.431437] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Lock "662d9fdf-ec0d-4466-8d15-abc942e4dad8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.433708] env[63372]: INFO nova.compute.manager [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Terminating instance [ 940.435273] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquiring lock "refresh_cache-662d9fdf-ec0d-4466-8d15-abc942e4dad8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.435436] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquired lock "refresh_cache-662d9fdf-ec0d-4466-8d15-abc942e4dad8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.435603] env[63372]: DEBUG nova.network.neutron [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 940.547634] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f3a3b36-5c42-4b8d-a3b9-a0484b047fc3 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.148s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.574912] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e2f6c7-ba69-228c-1049-37c4964b5069, 'name': SearchDatastore_Task, 'duration_secs': 0.010825} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.575199] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.575455] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] ed2cefa1-3604-43c5-9860-0fb49133bbb4/ed2cefa1-3604-43c5-9860-0fb49133bbb4.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 940.575713] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-09786364-4c72-4806-af92-3e1af4ceafa4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.582175] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Waiting for the task: (returnval){ [ 940.582175] env[63372]: value = "task-1024491" [ 940.582175] env[63372]: _type = "Task" [ 940.582175] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.590169] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024491, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.607048] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024490, 'name': ReconfigVM_Task, 'duration_secs': 1.025585} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.607048] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 91985614-b959-401e-bb06-d67b230ee026/91985614-b959-401e-bb06-d67b230ee026.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.607532] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4d2b5b87-532a-4b9f-b8e3-ecc4c89b23d5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.613219] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 940.613219] env[63372]: value = "task-1024492" [ 940.613219] env[63372]: _type = "Task" [ 940.613219] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.621775] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024492, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.955604] env[63372]: DEBUG nova.network.neutron [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 940.968382] env[63372]: DEBUG nova.network.neutron [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 941.021192] env[63372]: DEBUG nova.network.neutron [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.093673] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024491, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507176} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.093944] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] ed2cefa1-3604-43c5-9860-0fb49133bbb4/ed2cefa1-3604-43c5-9860-0fb49133bbb4.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 941.094193] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 941.094422] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2d2a4a09-f317-42e4-932f-03a16d4ce3e2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.103544] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Waiting for the task: (returnval){ [ 941.103544] env[63372]: value = "task-1024493" [ 941.103544] env[63372]: _type = "Task" [ 941.103544] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.117170] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024493, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.129310] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024492, 'name': Rename_Task, 'duration_secs': 0.163711} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.129661] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 941.129894] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f46c82d-61f2-4858-af3e-0e409470bf0e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.137027] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 941.137027] env[63372]: value = "task-1024494" [ 941.137027] env[63372]: _type = "Task" [ 941.137027] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.144303] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024494, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.173043] env[63372]: DEBUG nova.network.neutron [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Updating instance_info_cache with network_info: [{"id": "a6afe51b-6c5f-4f54-9657-97c43e85cdba", "address": "fa:16:3e:47:d5:c0", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6afe51b-6c", "ovs_interfaceid": "a6afe51b-6c5f-4f54-9657-97c43e85cdba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.233410] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a2b4da-f9fc-4739-b11a-b926b38275ba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.242011] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a46fdb-2b8f-4f95-8a2f-e9345d8859c8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.272883] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3883c2-d78f-4b18-abc0-d3e5259e78c6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.279349] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f727622-4ed0-4b65-89eb-547da6ef7743 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.294936] env[63372]: DEBUG nova.compute.provider_tree [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.524930] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Releasing lock "refresh_cache-662d9fdf-ec0d-4466-8d15-abc942e4dad8" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.525518] env[63372]: DEBUG nova.compute.manager [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 941.525793] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 941.526728] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855b5cc4-035a-42f0-956b-7e69bde28746 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.535931] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 941.536323] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dfb856ed-36dc-4d22-98a4-cef3233095e7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.543170] env[63372]: DEBUG oslo_vmware.api [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 941.543170] env[63372]: value = "task-1024495" [ 941.543170] env[63372]: _type = "Task" [ 941.543170] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.551614] env[63372]: DEBUG oslo_vmware.api [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024495, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.612972] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024493, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05884} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.613297] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.614092] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab684b9-c39f-4d9b-814c-dcb78989d884 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.636251] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] ed2cefa1-3604-43c5-9860-0fb49133bbb4/ed2cefa1-3604-43c5-9860-0fb49133bbb4.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.636571] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b4fea56-f414-4c3e-a0da-2d7230c3be56 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.659670] env[63372]: DEBUG oslo_vmware.api [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024494, 'name': PowerOnVM_Task, 'duration_secs': 0.455132} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.661048] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 941.661180] env[63372]: INFO nova.compute.manager [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Took 9.01 seconds to spawn the instance on the hypervisor. [ 941.661374] env[63372]: DEBUG nova.compute.manager [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.661749] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Waiting for the task: (returnval){ [ 941.661749] env[63372]: value = "task-1024496" [ 941.661749] env[63372]: _type = "Task" [ 941.661749] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.662527] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fba4f3c-3e5f-406a-aabc-a97cb8acd042 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.677750] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024496, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.677750] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "refresh_cache-4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.678292] env[63372]: DEBUG nova.compute.manager [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Instance network_info: |[{"id": "a6afe51b-6c5f-4f54-9657-97c43e85cdba", "address": "fa:16:3e:47:d5:c0", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6afe51b-6c", "ovs_interfaceid": "a6afe51b-6c5f-4f54-9657-97c43e85cdba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 941.678441] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:d5:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b21ab10-d886-4453-9472-9e11fb3c450d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a6afe51b-6c5f-4f54-9657-97c43e85cdba', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 941.686457] env[63372]: DEBUG oslo.service.loopingcall [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.686749] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 941.686965] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bbb7d2ce-f441-447f-9e06-3aaafcd1be48 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.708374] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 941.708374] env[63372]: value = "task-1024497" [ 941.708374] env[63372]: _type = "Task" [ 941.708374] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.718587] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024497, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.777112] env[63372]: DEBUG nova.compute.manager [req-80a18b6a-1801-40bc-8f3a-e1e813caeeec req-9e540890-a091-4100-bcc6-0e98e48ff8f0 service nova] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Received event network-changed-a6afe51b-6c5f-4f54-9657-97c43e85cdba {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.777265] env[63372]: DEBUG nova.compute.manager [req-80a18b6a-1801-40bc-8f3a-e1e813caeeec req-9e540890-a091-4100-bcc6-0e98e48ff8f0 service nova] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Refreshing instance network info cache due to event network-changed-a6afe51b-6c5f-4f54-9657-97c43e85cdba. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 941.777484] env[63372]: DEBUG oslo_concurrency.lockutils [req-80a18b6a-1801-40bc-8f3a-e1e813caeeec req-9e540890-a091-4100-bcc6-0e98e48ff8f0 service nova] Acquiring lock "refresh_cache-4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.777638] env[63372]: DEBUG oslo_concurrency.lockutils [req-80a18b6a-1801-40bc-8f3a-e1e813caeeec req-9e540890-a091-4100-bcc6-0e98e48ff8f0 service nova] Acquired lock "refresh_cache-4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.777821] env[63372]: DEBUG nova.network.neutron [req-80a18b6a-1801-40bc-8f3a-e1e813caeeec req-9e540890-a091-4100-bcc6-0e98e48ff8f0 service nova] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Refreshing network info cache for port a6afe51b-6c5f-4f54-9657-97c43e85cdba {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 941.798089] env[63372]: DEBUG nova.scheduler.client.report [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.054612] env[63372]: DEBUG oslo_vmware.api [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024495, 'name': PowerOffVM_Task, 'duration_secs': 0.129089} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.054937] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 942.055114] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 942.055376] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-82267d0b-1fe3-48d7-9012-6f7a06a51673 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.080162] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 942.080394] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 942.080582] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Deleting the datastore file [datastore1] 662d9fdf-ec0d-4466-8d15-abc942e4dad8 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.080893] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b15a5033-b735-4a13-996c-7ba992d1ed0a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.088412] env[63372]: DEBUG oslo_vmware.api [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for the task: (returnval){ [ 942.088412] env[63372]: value = "task-1024499" [ 942.088412] env[63372]: _type = "Task" [ 942.088412] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.096662] env[63372]: DEBUG oslo_vmware.api [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024499, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.116124] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "b7732621-7e58-40a3-b723-5c66df6f74a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.116371] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "b7732621-7e58-40a3-b723-5c66df6f74a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.175995] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024496, 'name': ReconfigVM_Task, 'duration_secs': 0.292151} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.176331] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Reconfigured VM instance instance-00000056 to attach disk [datastore1] ed2cefa1-3604-43c5-9860-0fb49133bbb4/ed2cefa1-3604-43c5-9860-0fb49133bbb4.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.176965] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de0d62ba-345c-488f-9f58-98cb70d4d12c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.185630] env[63372]: INFO nova.compute.manager [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Took 24.63 seconds to build instance. [ 942.192327] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Waiting for the task: (returnval){ [ 942.192327] env[63372]: value = "task-1024500" [ 942.192327] env[63372]: _type = "Task" [ 942.192327] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.205311] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024500, 'name': Rename_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.220749] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024497, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.303097] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.280s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.304277] env[63372]: DEBUG nova.compute.manager [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 942.307860] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.987s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.312020] env[63372]: INFO nova.compute.claims [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 942.543155] env[63372]: DEBUG nova.network.neutron [req-80a18b6a-1801-40bc-8f3a-e1e813caeeec req-9e540890-a091-4100-bcc6-0e98e48ff8f0 service nova] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Updated VIF entry in instance network info cache for port a6afe51b-6c5f-4f54-9657-97c43e85cdba. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 942.543534] env[63372]: DEBUG nova.network.neutron [req-80a18b6a-1801-40bc-8f3a-e1e813caeeec req-9e540890-a091-4100-bcc6-0e98e48ff8f0 service nova] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Updating instance_info_cache with network_info: [{"id": "a6afe51b-6c5f-4f54-9657-97c43e85cdba", "address": "fa:16:3e:47:d5:c0", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6afe51b-6c", "ovs_interfaceid": "a6afe51b-6c5f-4f54-9657-97c43e85cdba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.598334] env[63372]: DEBUG oslo_vmware.api [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Task: {'id': task-1024499, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.225419} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.598594] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 942.598782] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 942.598962] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 942.599152] env[63372]: INFO nova.compute.manager [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Took 1.07 seconds to destroy the instance on the hypervisor. [ 942.599391] env[63372]: DEBUG oslo.service.loopingcall [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.599582] env[63372]: DEBUG nova.compute.manager [-] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 942.599696] env[63372]: DEBUG nova.network.neutron [-] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 942.617106] env[63372]: DEBUG nova.network.neutron [-] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 942.618584] env[63372]: DEBUG nova.compute.manager [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 942.693789] env[63372]: DEBUG oslo_concurrency.lockutils [None req-45f60643-bb33-48d5-b8e8-7201d133e0a2 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "91985614-b959-401e-bb06-d67b230ee026" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.151s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.703115] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024500, 'name': Rename_Task, 'duration_secs': 0.142125} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.703388] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 942.705048] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e3adc70-98f6-4377-beba-4e3db09cc063 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.713917] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Waiting for the task: (returnval){ [ 942.713917] env[63372]: value = "task-1024501" [ 942.713917] env[63372]: _type = "Task" [ 942.713917] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.720344] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024497, 'name': CreateVM_Task, 'duration_secs': 0.623925} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.720804] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 942.721502] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.721672] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.722051] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 942.725147] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07a7dda8-e0b6-486d-8a3c-aee2ee3c92d6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.726441] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024501, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.729432] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 942.729432] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52dd4b39-959e-9124-f90c-f2a9052412d5" [ 942.729432] env[63372]: _type = "Task" [ 942.729432] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.740861] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52dd4b39-959e-9124-f90c-f2a9052412d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.818052] env[63372]: DEBUG nova.compute.utils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 942.821058] env[63372]: DEBUG nova.compute.manager [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 942.821058] env[63372]: DEBUG nova.network.neutron [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 942.860265] env[63372]: DEBUG nova.policy [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2253a9a409d4f1ba9176485655c0988', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dac773c2536745e285181426ae34bb96', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 942.961455] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "6beadbab-2cc1-4b69-95c9-e1a0ea11045f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.961706] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "6beadbab-2cc1-4b69-95c9-e1a0ea11045f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.047020] env[63372]: DEBUG oslo_concurrency.lockutils [req-80a18b6a-1801-40bc-8f3a-e1e813caeeec req-9e540890-a091-4100-bcc6-0e98e48ff8f0 service nova] Releasing lock "refresh_cache-4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.122623] env[63372]: DEBUG nova.network.neutron [-] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.152008] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.188459] env[63372]: DEBUG nova.network.neutron [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Successfully created port: 46264f2d-4aad-453f-b549-c48687e3b6ad {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 943.229517] env[63372]: DEBUG oslo_vmware.api [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024501, 'name': PowerOnVM_Task, 'duration_secs': 0.45165} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.229915] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 943.231975] env[63372]: INFO nova.compute.manager [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Took 7.27 seconds to spawn the instance on the hypervisor. [ 943.232308] env[63372]: DEBUG nova.compute.manager [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.236302] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d78c2799-6d3d-430f-a15b-3f1724932481 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.246712] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52dd4b39-959e-9124-f90c-f2a9052412d5, 'name': SearchDatastore_Task, 'duration_secs': 0.013013} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.248482] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.248823] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.249150] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.249370] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.249683] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.253166] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d65bfe53-819e-4f98-98d4-ae1806b8e5b7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.263510] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.263793] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 943.264754] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d23040c-c16f-44ac-b2a1-59b394961fff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.270625] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 943.270625] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525ca372-67ab-a581-f0ad-076e78e2dce9" [ 943.270625] env[63372]: _type = "Task" [ 943.270625] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.282027] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525ca372-67ab-a581-f0ad-076e78e2dce9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.321955] env[63372]: DEBUG nova.compute.manager [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 943.464319] env[63372]: DEBUG nova.compute.manager [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 943.572980] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf49086-4d75-4598-9d52-6efa1716e047 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.581210] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe9aabb-4708-4ac0-9840-9acb0c8e323c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.613709] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49553c7f-8581-45da-bf37-62d92370a65c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.621150] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f925f43-ee56-4efa-8e3e-26a8010c955f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.634510] env[63372]: INFO nova.compute.manager [-] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Took 1.03 seconds to deallocate network for instance. [ 943.635049] env[63372]: DEBUG nova.compute.provider_tree [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.762736] env[63372]: INFO nova.compute.manager [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Took 23.80 seconds to build instance. [ 943.781803] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525ca372-67ab-a581-f0ad-076e78e2dce9, 'name': SearchDatastore_Task, 'duration_secs': 0.012779} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.782643] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df14bcaa-283a-425e-804f-526be12b7638 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.788383] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 943.788383] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52a69b3d-44cd-e8c0-d67f-540a4b2c6896" [ 943.788383] env[63372]: _type = "Task" [ 943.788383] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.796297] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a69b3d-44cd-e8c0-d67f-540a4b2c6896, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.983128] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.139580] env[63372]: DEBUG nova.scheduler.client.report [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.143462] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.265789] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a71cad84-a1b3-4667-8666-6a13242164df tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Lock "ed2cefa1-3604-43c5-9860-0fb49133bbb4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.316s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.299850] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a69b3d-44cd-e8c0-d67f-540a4b2c6896, 'name': SearchDatastore_Task, 'duration_secs': 0.036984} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.301705] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.301705] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912/4ee6c0ac-b7ac-49e7-b029-ca6e9a595912.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 944.301705] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d4affa7a-8a2e-4138-b031-9a1275446704 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.307164] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 944.307164] env[63372]: value = "task-1024502" [ 944.307164] env[63372]: _type = "Task" [ 944.307164] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.315359] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024502, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.337975] env[63372]: DEBUG nova.compute.manager [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 944.363811] env[63372]: DEBUG nova.virt.hardware [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 944.364373] env[63372]: DEBUG nova.virt.hardware [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 944.364538] env[63372]: DEBUG nova.virt.hardware [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 944.364733] env[63372]: DEBUG nova.virt.hardware [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 944.364897] env[63372]: DEBUG nova.virt.hardware [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 944.365062] env[63372]: DEBUG nova.virt.hardware [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 944.365289] env[63372]: DEBUG nova.virt.hardware [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 944.365464] env[63372]: DEBUG nova.virt.hardware [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 944.365646] env[63372]: DEBUG nova.virt.hardware [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 944.365813] env[63372]: DEBUG nova.virt.hardware [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 944.366015] env[63372]: DEBUG nova.virt.hardware [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 944.367034] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ecf256-efa7-4f7e-8a11-65d79f0d4e1a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.376383] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da8c7d4-ac33-4adc-a68e-3982dffb8166 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.647394] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.648874] env[63372]: DEBUG oslo_concurrency.lockutils [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.294s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.649853] env[63372]: DEBUG nova.objects.instance [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lazy-loading 'resources' on Instance uuid 326a9155-1c00-4c18-a9a7-6124e709c20b {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.824223] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024502, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.873755] env[63372]: DEBUG nova.compute.manager [req-7ae2e3df-c4cf-4990-a897-9d2f4c979db8 req-63670e84-4b3a-4342-8651-3fe0f8a9a31f service nova] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Received event network-vif-plugged-46264f2d-4aad-453f-b549-c48687e3b6ad {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.873992] env[63372]: DEBUG oslo_concurrency.lockutils [req-7ae2e3df-c4cf-4990-a897-9d2f4c979db8 req-63670e84-4b3a-4342-8651-3fe0f8a9a31f service nova] Acquiring lock "61f7a052-9a42-4df3-a568-eab197fc5c4f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.874221] env[63372]: DEBUG oslo_concurrency.lockutils [req-7ae2e3df-c4cf-4990-a897-9d2f4c979db8 req-63670e84-4b3a-4342-8651-3fe0f8a9a31f service nova] Lock "61f7a052-9a42-4df3-a568-eab197fc5c4f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.874438] env[63372]: DEBUG oslo_concurrency.lockutils [req-7ae2e3df-c4cf-4990-a897-9d2f4c979db8 req-63670e84-4b3a-4342-8651-3fe0f8a9a31f service nova] Lock "61f7a052-9a42-4df3-a568-eab197fc5c4f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.874675] env[63372]: DEBUG nova.compute.manager [req-7ae2e3df-c4cf-4990-a897-9d2f4c979db8 req-63670e84-4b3a-4342-8651-3fe0f8a9a31f service nova] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] No waiting events found dispatching network-vif-plugged-46264f2d-4aad-453f-b549-c48687e3b6ad {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 944.874850] env[63372]: WARNING nova.compute.manager [req-7ae2e3df-c4cf-4990-a897-9d2f4c979db8 req-63670e84-4b3a-4342-8651-3fe0f8a9a31f service nova] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Received unexpected event network-vif-plugged-46264f2d-4aad-453f-b549-c48687e3b6ad for instance with vm_state building and task_state spawning. [ 944.981693] env[63372]: DEBUG nova.network.neutron [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Successfully updated port: 46264f2d-4aad-453f-b549-c48687e3b6ad {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 945.151786] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Acquiring lock "5ac4b598-1bef-46bc-8015-e345df861111" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.152122] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Lock "5ac4b598-1bef-46bc-8015-e345df861111" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.317677] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024502, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.763381} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.319893] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912/4ee6c0ac-b7ac-49e7-b029-ca6e9a595912.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 945.320134] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.320545] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9a02bed9-f6f2-497f-8e89-f1f110fc5d56 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.327119] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 945.327119] env[63372]: value = "task-1024503" [ 945.327119] env[63372]: _type = "Task" [ 945.327119] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.343537] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024503, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.355968] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1cc113-98f4-4173-b1d2-147e1c47bf30 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.363343] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01bfce3d-dfab-4f19-bf66-946373d5e870 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.395516] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c51a709-99c3-4384-b76b-aaa5f2c36d66 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.402413] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60da27e6-40e9-4962-9bbe-6d9e30c4dc2e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.415216] env[63372]: DEBUG nova.compute.provider_tree [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.484354] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "refresh_cache-61f7a052-9a42-4df3-a568-eab197fc5c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.484496] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "refresh_cache-61f7a052-9a42-4df3-a568-eab197fc5c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.484647] env[63372]: DEBUG nova.network.neutron [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 945.624660] env[63372]: DEBUG oslo_concurrency.lockutils [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Acquiring lock "ed2cefa1-3604-43c5-9860-0fb49133bbb4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.624934] env[63372]: DEBUG oslo_concurrency.lockutils [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Lock "ed2cefa1-3604-43c5-9860-0fb49133bbb4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.625168] env[63372]: DEBUG oslo_concurrency.lockutils [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Acquiring lock "ed2cefa1-3604-43c5-9860-0fb49133bbb4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.625360] env[63372]: DEBUG oslo_concurrency.lockutils [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Lock "ed2cefa1-3604-43c5-9860-0fb49133bbb4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.625534] env[63372]: DEBUG oslo_concurrency.lockutils [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Lock "ed2cefa1-3604-43c5-9860-0fb49133bbb4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.627795] env[63372]: INFO nova.compute.manager [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Terminating instance [ 945.629574] env[63372]: DEBUG nova.compute.manager [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 945.629825] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 945.630654] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966222b1-89f1-4d64-8244-86c28074f4fa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.638077] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 945.638077] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-485ce5fb-7c81-448e-ad2e-b2a450bdb5cc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.643804] env[63372]: DEBUG oslo_vmware.api [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Waiting for the task: (returnval){ [ 945.643804] env[63372]: value = "task-1024504" [ 945.643804] env[63372]: _type = "Task" [ 945.643804] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.652054] env[63372]: DEBUG oslo_vmware.api [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024504, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.656571] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Lock "5ac4b598-1bef-46bc-8015-e345df861111" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.504s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.657132] env[63372]: DEBUG nova.compute.manager [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 945.837122] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024503, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.382516} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.837505] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.838349] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db15df16-4e01-4610-ad8e-f13626668b11 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.861152] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912/4ee6c0ac-b7ac-49e7-b029-ca6e9a595912.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.861441] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7560299e-bf81-4762-a93e-a5a65e695761 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.879989] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 945.879989] env[63372]: value = "task-1024505" [ 945.879989] env[63372]: _type = "Task" [ 945.879989] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.887831] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024505, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.919031] env[63372]: DEBUG nova.scheduler.client.report [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.015937] env[63372]: DEBUG nova.network.neutron [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 946.153059] env[63372]: DEBUG oslo_vmware.api [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024504, 'name': PowerOffVM_Task, 'duration_secs': 0.198853} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.153928] env[63372]: DEBUG nova.network.neutron [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Updating instance_info_cache with network_info: [{"id": "46264f2d-4aad-453f-b549-c48687e3b6ad", "address": "fa:16:3e:c3:62:9c", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46264f2d-4a", "ovs_interfaceid": "46264f2d-4aad-453f-b549-c48687e3b6ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.155114] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 946.155296] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 946.155763] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6af04499-fa26-46ae-a47b-e435bc3bf98a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.161544] env[63372]: DEBUG nova.compute.utils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 946.162761] env[63372]: DEBUG nova.compute.manager [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 946.162931] env[63372]: DEBUG nova.network.neutron [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 946.200858] env[63372]: DEBUG nova.policy [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '45c33e156e444988b916c397e5de97eb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fde00a5b63714dd0975347bb573c1599', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 946.221800] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 946.222024] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 946.222221] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Deleting the datastore file [datastore1] ed2cefa1-3604-43c5-9860-0fb49133bbb4 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.222483] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b1343c37-fd84-421f-8e1d-25b8cb900491 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.228917] env[63372]: DEBUG oslo_vmware.api [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Waiting for the task: (returnval){ [ 946.228917] env[63372]: value = "task-1024507" [ 946.228917] env[63372]: _type = "Task" [ 946.228917] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.236481] env[63372]: DEBUG oslo_vmware.api [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024507, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.389740] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.425059] env[63372]: DEBUG oslo_concurrency.lockutils [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.776s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.427282] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.112s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.427545] env[63372]: DEBUG nova.objects.instance [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'resources' on Instance uuid 13bde897-8446-42a2-b02d-2f5b48e6f432 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.452876] env[63372]: INFO nova.scheduler.client.report [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleted allocations for instance 326a9155-1c00-4c18-a9a7-6124e709c20b [ 946.571146] env[63372]: DEBUG nova.network.neutron [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Successfully created port: eeee4694-73df-47f5-bd8c-82b44acebe2f {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 946.657171] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "refresh_cache-61f7a052-9a42-4df3-a568-eab197fc5c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.657510] env[63372]: DEBUG nova.compute.manager [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Instance network_info: |[{"id": "46264f2d-4aad-453f-b549-c48687e3b6ad", "address": "fa:16:3e:c3:62:9c", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46264f2d-4a", "ovs_interfaceid": "46264f2d-4aad-453f-b549-c48687e3b6ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 946.658105] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:62:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d4ef133-b6f3-41d1-add4-92a1482195cf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '46264f2d-4aad-453f-b549-c48687e3b6ad', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.668714] env[63372]: DEBUG oslo.service.loopingcall [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.669312] env[63372]: DEBUG nova.compute.manager [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 946.671955] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 946.675458] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3328a0e5-1cb8-40c8-acd8-25131c464b13 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.693355] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 946.693355] env[63372]: value = "task-1024508" [ 946.693355] env[63372]: _type = "Task" [ 946.693355] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.701253] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024508, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.739333] env[63372]: DEBUG oslo_vmware.api [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024507, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.891015] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024505, 'name': ReconfigVM_Task, 'duration_secs': 0.958965} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.891412] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912/4ee6c0ac-b7ac-49e7-b029-ca6e9a595912.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.891990] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c60d6339-1d05-45fc-b262-fe3124428c0c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.898222] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 946.898222] env[63372]: value = "task-1024509" [ 946.898222] env[63372]: _type = "Task" [ 946.898222] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.900162] env[63372]: DEBUG nova.compute.manager [req-621a1602-1365-45eb-9992-6d2827274bfa req-353c4eee-423f-4328-8644-550cd8e4beda service nova] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Received event network-changed-46264f2d-4aad-453f-b549-c48687e3b6ad {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.901097] env[63372]: DEBUG nova.compute.manager [req-621a1602-1365-45eb-9992-6d2827274bfa req-353c4eee-423f-4328-8644-550cd8e4beda service nova] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Refreshing instance network info cache due to event network-changed-46264f2d-4aad-453f-b549-c48687e3b6ad. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 946.901097] env[63372]: DEBUG oslo_concurrency.lockutils [req-621a1602-1365-45eb-9992-6d2827274bfa req-353c4eee-423f-4328-8644-550cd8e4beda service nova] Acquiring lock "refresh_cache-61f7a052-9a42-4df3-a568-eab197fc5c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.901097] env[63372]: DEBUG oslo_concurrency.lockutils [req-621a1602-1365-45eb-9992-6d2827274bfa req-353c4eee-423f-4328-8644-550cd8e4beda service nova] Acquired lock "refresh_cache-61f7a052-9a42-4df3-a568-eab197fc5c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.901097] env[63372]: DEBUG nova.network.neutron [req-621a1602-1365-45eb-9992-6d2827274bfa req-353c4eee-423f-4328-8644-550cd8e4beda service nova] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Refreshing network info cache for port 46264f2d-4aad-453f-b549-c48687e3b6ad {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 946.911557] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024509, 'name': Rename_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.962476] env[63372]: DEBUG oslo_concurrency.lockutils [None req-77b7cab5-dc27-4f31-a68f-5ade447bac57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "326a9155-1c00-4c18-a9a7-6124e709c20b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.627s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.115043] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a87eef-2f19-4aba-b848-7d6a0a39b869 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.121788] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0cc43f-930c-4070-bdff-358ceefa8820 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.151693] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f93bf00-937c-41ec-87f7-921bd58561da {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.159020] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11daa07-6249-484f-a96a-b0066b86f177 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.172353] env[63372]: DEBUG nova.compute.provider_tree [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 947.202030] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024508, 'name': CreateVM_Task, 'duration_secs': 0.32879} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.202198] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 947.202864] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.203043] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.203357] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 947.203607] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b699207f-42e0-4e67-9630-c9f7bb3e8249 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.208493] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 947.208493] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52291875-46df-6bbe-812c-23481f1d92fb" [ 947.208493] env[63372]: _type = "Task" [ 947.208493] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.216229] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52291875-46df-6bbe-812c-23481f1d92fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.237608] env[63372]: DEBUG oslo_vmware.api [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Task: {'id': task-1024507, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.515601} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.237855] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 947.238069] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 947.238257] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 947.238444] env[63372]: INFO nova.compute.manager [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Took 1.61 seconds to destroy the instance on the hypervisor. [ 947.238678] env[63372]: DEBUG oslo.service.loopingcall [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.238865] env[63372]: DEBUG nova.compute.manager [-] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 947.239040] env[63372]: DEBUG nova.network.neutron [-] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 947.411393] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024509, 'name': Rename_Task, 'duration_secs': 0.16194} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.411963] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 947.412233] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-84b3ce30-7e83-415f-ba48-02ffc87f99b9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.418269] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 947.418269] env[63372]: value = "task-1024510" [ 947.418269] env[63372]: _type = "Task" [ 947.418269] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.425743] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024510, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.681384] env[63372]: DEBUG nova.compute.manager [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 947.712978] env[63372]: ERROR nova.scheduler.client.report [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [req-4ba2cf90-5511-4923-b420-a4a69089b858] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4ba2cf90-5511-4923-b420-a4a69089b858"}]} [ 947.724945] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52291875-46df-6bbe-812c-23481f1d92fb, 'name': SearchDatastore_Task, 'duration_secs': 0.00987} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.730715] env[63372]: DEBUG nova.virt.hardware [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 947.730981] env[63372]: DEBUG nova.virt.hardware [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 947.731168] env[63372]: DEBUG nova.virt.hardware [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 947.731357] env[63372]: DEBUG nova.virt.hardware [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 947.731505] env[63372]: DEBUG nova.virt.hardware [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 947.731654] env[63372]: DEBUG nova.virt.hardware [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 947.731859] env[63372]: DEBUG nova.virt.hardware [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 947.733136] env[63372]: DEBUG nova.virt.hardware [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 947.733381] env[63372]: DEBUG nova.virt.hardware [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 947.733566] env[63372]: DEBUG nova.virt.hardware [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 947.733745] env[63372]: DEBUG nova.virt.hardware [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 947.734092] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.735717] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.737093] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.737717] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.737717] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.738669] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f5cab15-08c9-402f-a6f0-96a21e1d0f92 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.741993] env[63372]: DEBUG nova.scheduler.client.report [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Refreshing inventories for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 947.743792] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-113d3c61-2724-4215-a475-7d131cbf8772 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.751847] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ee35d6-e5f0-4b16-b8ee-98ab454d50e9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.757448] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.757637] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 947.758677] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b09e89ad-b283-4445-9bd5-be1c105f9160 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.769300] env[63372]: DEBUG nova.scheduler.client.report [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Updating ProviderTree inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 947.769514] env[63372]: DEBUG nova.compute.provider_tree [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 947.776701] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 947.776701] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]527e658f-9cd3-9cd6-8898-97b814833545" [ 947.776701] env[63372]: _type = "Task" [ 947.776701] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.784895] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527e658f-9cd3-9cd6-8898-97b814833545, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.786545] env[63372]: DEBUG nova.scheduler.client.report [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Refreshing aggregate associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, aggregates: None {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 947.804546] env[63372]: DEBUG nova.scheduler.client.report [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Refreshing trait associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 947.865744] env[63372]: DEBUG nova.compute.manager [req-37de0b3c-3491-4473-86ba-b5ee63fa1519 req-7880327e-c5c1-4fde-a7e8-2dc3e37ab820 service nova] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Received event network-vif-deleted-e129d65d-67e7-49fb-95d6-2b57141bd3b3 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.865955] env[63372]: INFO nova.compute.manager [req-37de0b3c-3491-4473-86ba-b5ee63fa1519 req-7880327e-c5c1-4fde-a7e8-2dc3e37ab820 service nova] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Neutron deleted interface e129d65d-67e7-49fb-95d6-2b57141bd3b3; detaching it from the instance and deleting it from the info cache [ 947.866149] env[63372]: DEBUG nova.network.neutron [req-37de0b3c-3491-4473-86ba-b5ee63fa1519 req-7880327e-c5c1-4fde-a7e8-2dc3e37ab820 service nova] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.934092] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024510, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.999357] env[63372]: DEBUG nova.network.neutron [-] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.020432] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a7000ce-e1d2-4b9a-b041-b5a645795bab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.030066] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662fefdb-e8e9-411e-b530-5c5fe5bd2123 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.067212] env[63372]: DEBUG nova.network.neutron [req-621a1602-1365-45eb-9992-6d2827274bfa req-353c4eee-423f-4328-8644-550cd8e4beda service nova] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Updated VIF entry in instance network info cache for port 46264f2d-4aad-453f-b549-c48687e3b6ad. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 948.068276] env[63372]: DEBUG nova.network.neutron [req-621a1602-1365-45eb-9992-6d2827274bfa req-353c4eee-423f-4328-8644-550cd8e4beda service nova] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Updating instance_info_cache with network_info: [{"id": "46264f2d-4aad-453f-b549-c48687e3b6ad", "address": "fa:16:3e:c3:62:9c", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46264f2d-4a", "ovs_interfaceid": "46264f2d-4aad-453f-b549-c48687e3b6ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.069448] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b151ce-95e7-4b3b-bbd7-8f3bd093b80c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.077793] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c4cea4-5bf4-46d6-8867-1a6b7e79ded6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.092852] env[63372]: DEBUG nova.compute.provider_tree [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 948.287535] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527e658f-9cd3-9cd6-8898-97b814833545, 'name': SearchDatastore_Task, 'duration_secs': 0.009775} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.288410] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86539b4d-3c6e-4d01-b3ac-d8349f97b375 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.293891] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 948.293891] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5226743c-4dc0-6dec-0a30-f3f857682837" [ 948.293891] env[63372]: _type = "Task" [ 948.293891] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.301790] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5226743c-4dc0-6dec-0a30-f3f857682837, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.369559] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e35d0cf5-7ed6-4cd0-ad4c-6c15af0b6991 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.379626] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac94aff-bb31-402c-bf0a-e26ed1802992 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.392641] env[63372]: DEBUG nova.network.neutron [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Successfully updated port: eeee4694-73df-47f5-bd8c-82b44acebe2f {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 948.413584] env[63372]: DEBUG nova.compute.manager [req-37de0b3c-3491-4473-86ba-b5ee63fa1519 req-7880327e-c5c1-4fde-a7e8-2dc3e37ab820 service nova] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Detach interface failed, port_id=e129d65d-67e7-49fb-95d6-2b57141bd3b3, reason: Instance ed2cefa1-3604-43c5-9860-0fb49133bbb4 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 948.431863] env[63372]: DEBUG oslo_vmware.api [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024510, 'name': PowerOnVM_Task, 'duration_secs': 0.742045} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.432147] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 948.432356] env[63372]: INFO nova.compute.manager [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Took 9.21 seconds to spawn the instance on the hypervisor. [ 948.432537] env[63372]: DEBUG nova.compute.manager [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.433324] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0216cc5-1f47-40ed-bce1-415641fe0830 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.505115] env[63372]: INFO nova.compute.manager [-] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Took 1.27 seconds to deallocate network for instance. [ 948.573224] env[63372]: DEBUG oslo_concurrency.lockutils [req-621a1602-1365-45eb-9992-6d2827274bfa req-353c4eee-423f-4328-8644-550cd8e4beda service nova] Releasing lock "refresh_cache-61f7a052-9a42-4df3-a568-eab197fc5c4f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.621226] env[63372]: DEBUG nova.scheduler.client.report [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 122 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 948.621508] env[63372]: DEBUG nova.compute.provider_tree [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 122 to 123 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 948.621688] env[63372]: DEBUG nova.compute.provider_tree [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 948.804717] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5226743c-4dc0-6dec-0a30-f3f857682837, 'name': SearchDatastore_Task, 'duration_secs': 0.037148} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.804934] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.805217] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 61f7a052-9a42-4df3-a568-eab197fc5c4f/61f7a052-9a42-4df3-a568-eab197fc5c4f.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 948.805477] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6b0d5097-63cd-411b-836b-45b8df401e02 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.812970] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 948.812970] env[63372]: value = "task-1024511" [ 948.812970] env[63372]: _type = "Task" [ 948.812970] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.820667] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024511, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.895916] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Acquiring lock "refresh_cache-c7c444cb-0aaa-4486-840a-e48fa255b421" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.896641] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Acquired lock "refresh_cache-c7c444cb-0aaa-4486-840a-e48fa255b421" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.896641] env[63372]: DEBUG nova.network.neutron [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 948.942283] env[63372]: DEBUG nova.compute.manager [req-a9884933-bec8-4a78-8c94-f4948782a564 req-57723d8f-9fe0-40f0-89fb-051e2c1b1b7f service nova] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Received event network-vif-plugged-eeee4694-73df-47f5-bd8c-82b44acebe2f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.942475] env[63372]: DEBUG oslo_concurrency.lockutils [req-a9884933-bec8-4a78-8c94-f4948782a564 req-57723d8f-9fe0-40f0-89fb-051e2c1b1b7f service nova] Acquiring lock "c7c444cb-0aaa-4486-840a-e48fa255b421-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.942624] env[63372]: DEBUG oslo_concurrency.lockutils [req-a9884933-bec8-4a78-8c94-f4948782a564 req-57723d8f-9fe0-40f0-89fb-051e2c1b1b7f service nova] Lock "c7c444cb-0aaa-4486-840a-e48fa255b421-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.942735] env[63372]: DEBUG oslo_concurrency.lockutils [req-a9884933-bec8-4a78-8c94-f4948782a564 req-57723d8f-9fe0-40f0-89fb-051e2c1b1b7f service nova] Lock "c7c444cb-0aaa-4486-840a-e48fa255b421-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.942891] env[63372]: DEBUG nova.compute.manager [req-a9884933-bec8-4a78-8c94-f4948782a564 req-57723d8f-9fe0-40f0-89fb-051e2c1b1b7f service nova] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] No waiting events found dispatching network-vif-plugged-eeee4694-73df-47f5-bd8c-82b44acebe2f {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 948.943072] env[63372]: WARNING nova.compute.manager [req-a9884933-bec8-4a78-8c94-f4948782a564 req-57723d8f-9fe0-40f0-89fb-051e2c1b1b7f service nova] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Received unexpected event network-vif-plugged-eeee4694-73df-47f5-bd8c-82b44acebe2f for instance with vm_state building and task_state spawning. [ 948.943244] env[63372]: DEBUG nova.compute.manager [req-a9884933-bec8-4a78-8c94-f4948782a564 req-57723d8f-9fe0-40f0-89fb-051e2c1b1b7f service nova] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Received event network-changed-eeee4694-73df-47f5-bd8c-82b44acebe2f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.943483] env[63372]: DEBUG nova.compute.manager [req-a9884933-bec8-4a78-8c94-f4948782a564 req-57723d8f-9fe0-40f0-89fb-051e2c1b1b7f service nova] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Refreshing instance network info cache due to event network-changed-eeee4694-73df-47f5-bd8c-82b44acebe2f. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 948.943546] env[63372]: DEBUG oslo_concurrency.lockutils [req-a9884933-bec8-4a78-8c94-f4948782a564 req-57723d8f-9fe0-40f0-89fb-051e2c1b1b7f service nova] Acquiring lock "refresh_cache-c7c444cb-0aaa-4486-840a-e48fa255b421" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.950867] env[63372]: INFO nova.compute.manager [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Took 24.45 seconds to build instance. [ 949.011922] env[63372]: DEBUG oslo_concurrency.lockutils [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.127234] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.700s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.129870] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 12.070s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.151022] env[63372]: INFO nova.scheduler.client.report [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Deleted allocations for instance 13bde897-8446-42a2-b02d-2f5b48e6f432 [ 949.324406] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024511, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473729} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.324909] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 61f7a052-9a42-4df3-a568-eab197fc5c4f/61f7a052-9a42-4df3-a568-eab197fc5c4f.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 949.324909] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 949.325262] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9168b572-25aa-4d27-bc4e-4823153961e9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.332388] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 949.332388] env[63372]: value = "task-1024512" [ 949.332388] env[63372]: _type = "Task" [ 949.332388] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.341443] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024512, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.438710] env[63372]: DEBUG nova.network.neutron [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 949.453583] env[63372]: DEBUG oslo_concurrency.lockutils [None req-17a50889-4e9f-4206-8630-20c3627a944a tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.969s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.577085] env[63372]: DEBUG nova.network.neutron [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Updating instance_info_cache with network_info: [{"id": "eeee4694-73df-47f5-bd8c-82b44acebe2f", "address": "fa:16:3e:69:de:1a", "network": {"id": "55aa0b02-1de9-4cbd-a11d-dd228d97fa9a", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-853194516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fde00a5b63714dd0975347bb573c1599", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeeee4694-73", "ovs_interfaceid": "eeee4694-73df-47f5-bd8c-82b44acebe2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.635535] env[63372]: INFO nova.compute.claims [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 949.659290] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9de083ad-076e-4271-9cba-10b97f6431f7 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "13bde897-8446-42a2-b02d-2f5b48e6f432" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.934s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.660709] env[63372]: DEBUG oslo_concurrency.lockutils [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] Acquired lock "13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.661578] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d31a409-9a2a-43d0-b708-ddfb538de534 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.671569] env[63372]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 949.671731] env[63372]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=63372) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 949.672389] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bebbaac1-6cc1-4f84-a580-1fd41caabe89 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.680790] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8a5052-27a1-464c-ac45-7164c2a5bbf4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.709423] env[63372]: ERROR root [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-227384' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-227384' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-227384' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-227384'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-227384' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-227384' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-227384'}\n"]: nova.exception.InstanceNotFound: Instance 13bde897-8446-42a2-b02d-2f5b48e6f432 could not be found. [ 949.709654] env[63372]: DEBUG oslo_concurrency.lockutils [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] Releasing lock "13bde897-8446-42a2-b02d-2f5b48e6f432" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.709958] env[63372]: DEBUG nova.compute.manager [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Detach interface failed, port_id=a5e3b219-68f3-4fd7-a08c-d9cf842a5e2c, reason: Instance 13bde897-8446-42a2-b02d-2f5b48e6f432 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 949.710196] env[63372]: DEBUG nova.compute.manager [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Received event network-vif-deleted-8321625e-abd2-4cc3-b545-2190d9759158 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.710373] env[63372]: INFO nova.compute.manager [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Neutron deleted interface 8321625e-abd2-4cc3-b545-2190d9759158; detaching it from the instance and deleting it from the info cache [ 949.710585] env[63372]: DEBUG nova.network.neutron [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.843035] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024512, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077937} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.843290] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 949.844073] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3679d010-57c5-4749-afa5-34d63e371300 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.865833] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 61f7a052-9a42-4df3-a568-eab197fc5c4f/61f7a052-9a42-4df3-a568-eab197fc5c4f.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.866116] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-848e2ac8-dbef-405b-996d-57b5d2a5db30 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.885018] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 949.885018] env[63372]: value = "task-1024513" [ 949.885018] env[63372]: _type = "Task" [ 949.885018] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.893280] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024513, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.079988] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Releasing lock "refresh_cache-c7c444cb-0aaa-4486-840a-e48fa255b421" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.080374] env[63372]: DEBUG nova.compute.manager [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Instance network_info: |[{"id": "eeee4694-73df-47f5-bd8c-82b44acebe2f", "address": "fa:16:3e:69:de:1a", "network": {"id": "55aa0b02-1de9-4cbd-a11d-dd228d97fa9a", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-853194516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fde00a5b63714dd0975347bb573c1599", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeeee4694-73", "ovs_interfaceid": "eeee4694-73df-47f5-bd8c-82b44acebe2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 950.080741] env[63372]: DEBUG oslo_concurrency.lockutils [req-a9884933-bec8-4a78-8c94-f4948782a564 req-57723d8f-9fe0-40f0-89fb-051e2c1b1b7f service nova] Acquired lock "refresh_cache-c7c444cb-0aaa-4486-840a-e48fa255b421" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.080974] env[63372]: DEBUG nova.network.neutron [req-a9884933-bec8-4a78-8c94-f4948782a564 req-57723d8f-9fe0-40f0-89fb-051e2c1b1b7f service nova] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Refreshing network info cache for port eeee4694-73df-47f5-bd8c-82b44acebe2f {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 950.082251] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:69:de:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '023d6500-887e-4dc4-bec5-06b40450d9c0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eeee4694-73df-47f5-bd8c-82b44acebe2f', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 950.090032] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Creating folder: Project (fde00a5b63714dd0975347bb573c1599). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 950.090596] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d65a3c10-0852-413b-aa72-93d4b80c55ab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.102049] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Created folder: Project (fde00a5b63714dd0975347bb573c1599) in parent group-v227230. [ 950.102208] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Creating folder: Instances. Parent ref: group-v227423. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 950.102459] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bb1c3b67-56a9-4c9e-8dff-0f46619a9ff4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.111279] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Created folder: Instances in parent group-v227423. [ 950.111551] env[63372]: DEBUG oslo.service.loopingcall [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.111752] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 950.111977] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-41f1dba0-4374-41a3-a253-381239146b9e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.130532] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 950.130532] env[63372]: value = "task-1024516" [ 950.130532] env[63372]: _type = "Task" [ 950.130532] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.138379] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024516, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.142388] env[63372]: INFO nova.compute.resource_tracker [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating resource usage from migration 4130afc1-bde5-42b0-a496-4513ec7d4f56 [ 950.213541] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff84730a-2af8-420f-a425-0152b6b2a00b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.222859] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf86f901-8556-459e-9948-864a3d5adbc9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.256147] env[63372]: DEBUG nova.compute.manager [req-273189e6-806f-4b53-890a-1df9037e6128 req-780b6bca-dba2-48b0-be24-d9a9e0293884 service nova] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Detach interface failed, port_id=8321625e-abd2-4cc3-b545-2190d9759158, reason: Instance 326a9155-1c00-4c18-a9a7-6124e709c20b could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 950.339309] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b74f4e-3c5b-43db-97a6-f7ee713e9847 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.348270] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16de9850-5604-4d32-97fe-95ae8a8987f4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.386332] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e60d51de-3bc1-4355-9343-a26616fab10f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.400254] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c847fcb-e75c-4f61-ae55-f61faf9919ed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.404684] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024513, 'name': ReconfigVM_Task, 'duration_secs': 0.308369} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.404967] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 61f7a052-9a42-4df3-a568-eab197fc5c4f/61f7a052-9a42-4df3-a568-eab197fc5c4f.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.405975] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d85556d2-11cc-442a-9afa-b3c89479c884 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.416307] env[63372]: DEBUG nova.compute.provider_tree [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.421934] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 950.421934] env[63372]: value = "task-1024517" [ 950.421934] env[63372]: _type = "Task" [ 950.421934] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.430335] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024517, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.641163] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024516, 'name': CreateVM_Task, 'duration_secs': 0.35944} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.641344] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 950.642008] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.642188] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.642493] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 950.644785] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47bbf16c-b9ee-4735-84bd-1e96e1cdcefd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.649685] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Waiting for the task: (returnval){ [ 950.649685] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521395ab-4075-c30a-9fc5-426cc2ae6b0f" [ 950.649685] env[63372]: _type = "Task" [ 950.649685] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.657333] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521395ab-4075-c30a-9fc5-426cc2ae6b0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.750056] env[63372]: DEBUG oslo_concurrency.lockutils [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.750320] env[63372]: DEBUG oslo_concurrency.lockutils [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.906612] env[63372]: DEBUG nova.network.neutron [req-a9884933-bec8-4a78-8c94-f4948782a564 req-57723d8f-9fe0-40f0-89fb-051e2c1b1b7f service nova] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Updated VIF entry in instance network info cache for port eeee4694-73df-47f5-bd8c-82b44acebe2f. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 950.906986] env[63372]: DEBUG nova.network.neutron [req-a9884933-bec8-4a78-8c94-f4948782a564 req-57723d8f-9fe0-40f0-89fb-051e2c1b1b7f service nova] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Updating instance_info_cache with network_info: [{"id": "eeee4694-73df-47f5-bd8c-82b44acebe2f", "address": "fa:16:3e:69:de:1a", "network": {"id": "55aa0b02-1de9-4cbd-a11d-dd228d97fa9a", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-853194516-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fde00a5b63714dd0975347bb573c1599", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeeee4694-73", "ovs_interfaceid": "eeee4694-73df-47f5-bd8c-82b44acebe2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.920647] env[63372]: DEBUG nova.scheduler.client.report [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.935263] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024517, 'name': Rename_Task, 'duration_secs': 0.150563} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.935389] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 950.935615] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3723b0f1-20c8-4e19-bcbf-b9c42ec4e749 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.942600] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 950.942600] env[63372]: value = "task-1024518" [ 950.942600] env[63372]: _type = "Task" [ 950.942600] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.951409] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024518, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.163022] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521395ab-4075-c30a-9fc5-426cc2ae6b0f, 'name': SearchDatastore_Task, 'duration_secs': 0.063427} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.163022] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.163022] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 951.163022] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.163022] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.163022] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 951.163022] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0f46c9a-9985-42eb-8bb8-3c71ce545e27 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.173377] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 951.173740] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 951.174647] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f805a06-07ab-4813-87de-076b8c5ef25e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.180998] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Waiting for the task: (returnval){ [ 951.180998] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f24d46-ea1b-57d2-b66e-d884deb5447f" [ 951.180998] env[63372]: _type = "Task" [ 951.180998] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.188880] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f24d46-ea1b-57d2-b66e-d884deb5447f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.254189] env[63372]: DEBUG nova.compute.utils [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 951.412710] env[63372]: DEBUG oslo_concurrency.lockutils [req-a9884933-bec8-4a78-8c94-f4948782a564 req-57723d8f-9fe0-40f0-89fb-051e2c1b1b7f service nova] Releasing lock "refresh_cache-c7c444cb-0aaa-4486-840a-e48fa255b421" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.425925] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.296s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.426185] env[63372]: INFO nova.compute.manager [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Migrating [ 951.434248] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.876s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.434510] env[63372]: DEBUG nova.objects.instance [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lazy-loading 'resources' on Instance uuid 95178e2e-9d71-4606-a856-c530bfbb9345 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.454719] env[63372]: DEBUG oslo_vmware.api [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024518, 'name': PowerOnVM_Task, 'duration_secs': 0.460936} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.454979] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 951.455202] env[63372]: INFO nova.compute.manager [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Took 7.12 seconds to spawn the instance on the hypervisor. [ 951.455380] env[63372]: DEBUG nova.compute.manager [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 951.456360] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae9b492-1aab-42d5-b018-440aa91e8fe9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.473293] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "b455b16b-3332-43bf-a91e-a350287ba5f3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.473525] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "b455b16b-3332-43bf-a91e-a350287ba5f3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.692188] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f24d46-ea1b-57d2-b66e-d884deb5447f, 'name': SearchDatastore_Task, 'duration_secs': 0.010673} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.693022] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40b9be0f-be4e-48b9-95c5-3dc6fa2aa4fc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.698249] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Waiting for the task: (returnval){ [ 951.698249] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52d35f77-081e-91d0-1c40-c5530bdb3abf" [ 951.698249] env[63372]: _type = "Task" [ 951.698249] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.707562] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52d35f77-081e-91d0-1c40-c5530bdb3abf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.755956] env[63372]: DEBUG oslo_concurrency.lockutils [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.944037] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.944141] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.944348] env[63372]: DEBUG nova.network.neutron [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 951.974629] env[63372]: INFO nova.compute.manager [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Took 23.20 seconds to build instance. [ 951.975589] env[63372]: DEBUG nova.compute.manager [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 952.110475] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97054e84-c4d8-4998-8543-441f8b6e8936 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.117859] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6401bd80-b94b-4a76-bba1-16a0cbe3c12f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.147656] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0adc36a0-d559-4686-a722-67146220d657 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.155039] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebebbd3e-ee27-4eb7-bed0-b4bc78040721 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.167911] env[63372]: DEBUG nova.compute.provider_tree [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.207494] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52d35f77-081e-91d0-1c40-c5530bdb3abf, 'name': SearchDatastore_Task, 'duration_secs': 0.019525} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.207776] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.208051] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] c7c444cb-0aaa-4486-840a-e48fa255b421/c7c444cb-0aaa-4486-840a-e48fa255b421.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 952.208531] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0a173a3a-f948-4d27-92a5-c6b1f8ed8a2b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.215518] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Waiting for the task: (returnval){ [ 952.215518] env[63372]: value = "task-1024519" [ 952.215518] env[63372]: _type = "Task" [ 952.215518] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.222815] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024519, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.479250] env[63372]: DEBUG oslo_concurrency.lockutils [None req-542387e9-1b0c-4bc5-8e20-5fc9730b8a47 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "61f7a052-9a42-4df3-a568-eab197fc5c4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.316s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.494998] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.670989] env[63372]: DEBUG nova.scheduler.client.report [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.727256] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024519, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.790031] env[63372]: DEBUG nova.network.neutron [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance_info_cache with network_info: [{"id": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "address": "fa:16:3e:3f:99:77", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c4c74ba-ad", "ovs_interfaceid": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.820395] env[63372]: DEBUG oslo_concurrency.lockutils [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.820662] env[63372]: DEBUG oslo_concurrency.lockutils [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.820921] env[63372]: INFO nova.compute.manager [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Attaching volume 3673857b-aabe-4a87-9f39-54d3076eee42 to /dev/sdb [ 952.861274] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c07883-94ea-4195-a547-a76c5ba36378 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.870439] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62aac095-ff1f-4b15-a9c9-4736ea314ad4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.885937] env[63372]: DEBUG nova.virt.block_device [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Updating existing volume attachment record: 59bbeaae-829d-4cb9-aa5f-11c08e15f5a0 {{(pid=63372) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 953.176599] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.742s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.180110] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 13.196s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.180495] env[63372]: DEBUG nova.objects.instance [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63372) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 953.197270] env[63372]: INFO nova.scheduler.client.report [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleted allocations for instance 95178e2e-9d71-4606-a856-c530bfbb9345 [ 953.228067] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024519, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.844547} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.228067] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] c7c444cb-0aaa-4486-840a-e48fa255b421/c7c444cb-0aaa-4486-840a-e48fa255b421.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 953.228295] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 953.228518] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-488cf74c-84eb-43a6-badb-c52aa1012b2f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.235738] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Waiting for the task: (returnval){ [ 953.235738] env[63372]: value = "task-1024523" [ 953.235738] env[63372]: _type = "Task" [ 953.235738] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.244874] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024523, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.293605] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.587113] env[63372]: INFO nova.compute.manager [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Rebuilding instance [ 953.623424] env[63372]: DEBUG nova.compute.manager [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.624298] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a25669-2778-4d4a-bc8d-a9bf37fd6ca1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.705162] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f2267e63-a1af-4a30-aa58-4eedc670ec6a tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "95178e2e-9d71-4606-a856-c530bfbb9345" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.072s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.744516] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024523, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061068} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.744762] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 953.745499] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8549da50-64ed-4c2f-8dc1-03ec3c25eec3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.766794] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] c7c444cb-0aaa-4486-840a-e48fa255b421/c7c444cb-0aaa-4486-840a-e48fa255b421.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 953.767019] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72bfb10f-c334-43ef-899e-2b841cdbd2f2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.785869] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Waiting for the task: (returnval){ [ 953.785869] env[63372]: value = "task-1024524" [ 953.785869] env[63372]: _type = "Task" [ 953.785869] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.793380] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024524, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.134282] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 954.134625] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-921f9b52-73e9-4c1a-9c10-1fdae186afa7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.142529] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 954.142529] env[63372]: value = "task-1024525" [ 954.142529] env[63372]: _type = "Task" [ 954.142529] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.152878] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024525, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.189579] env[63372]: DEBUG oslo_concurrency.lockutils [None req-46f4bf51-7ca3-4036-9a4c-8b59aec62d29 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.191367] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.040s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.193314] env[63372]: INFO nova.compute.claims [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 954.296326] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024524, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.428710] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "ba1daa0a-dcf8-4586-9964-a962fcc94196" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.429008] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "ba1daa0a-dcf8-4586-9964-a962fcc94196" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.429245] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "ba1daa0a-dcf8-4586-9964-a962fcc94196-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.429434] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "ba1daa0a-dcf8-4586-9964-a962fcc94196-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.429606] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "ba1daa0a-dcf8-4586-9964-a962fcc94196-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.432206] env[63372]: INFO nova.compute.manager [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Terminating instance [ 954.433952] env[63372]: DEBUG nova.compute.manager [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 954.434161] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 954.434975] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f8fb4e-6406-4f86-9b64-6123c8681205 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.442319] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 954.442540] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6937f49b-d448-45ae-b67f-cf2b797fdbee {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.448257] env[63372]: DEBUG oslo_vmware.api [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 954.448257] env[63372]: value = "task-1024526" [ 954.448257] env[63372]: _type = "Task" [ 954.448257] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.455433] env[63372]: DEBUG oslo_vmware.api [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024526, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.654811] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024525, 'name': PowerOffVM_Task, 'duration_secs': 0.233434} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.655207] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 954.655322] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 954.656158] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f24eb5-c536-4e0b-a029-53979f2ece58 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.662512] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 954.662734] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f23ee923-5a91-4eb0-92a3-805c15721967 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.730412] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 954.730625] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 954.730810] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleting the datastore file [datastore2] 61f7a052-9a42-4df3-a568-eab197fc5c4f {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.731143] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6e868eec-2107-49b9-987a-ad7d30ba5439 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.738083] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 954.738083] env[63372]: value = "task-1024528" [ 954.738083] env[63372]: _type = "Task" [ 954.738083] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.746124] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024528, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.796646] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024524, 'name': ReconfigVM_Task, 'duration_secs': 0.570778} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.796929] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Reconfigured VM instance instance-00000059 to attach disk [datastore2] c7c444cb-0aaa-4486-840a-e48fa255b421/c7c444cb-0aaa-4486-840a-e48fa255b421.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 954.797553] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cffc7daf-c49e-4d28-b405-65bccbbd3d2d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.805790] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Waiting for the task: (returnval){ [ 954.805790] env[63372]: value = "task-1024529" [ 954.805790] env[63372]: _type = "Task" [ 954.805790] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.806902] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47e3960-f19f-4b32-8c16-c7f610dc0851 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.816709] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024529, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.829830] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance '93a5d948-0629-4f53-a681-858d519acfa7' progress to 0 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 954.958030] env[63372]: DEBUG oslo_vmware.api [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024526, 'name': PowerOffVM_Task, 'duration_secs': 0.473526} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.958154] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 954.958315] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 954.958603] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a82585ba-9c07-47fb-9b73-6423cf332e4a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.046184] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 955.046421] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 955.046615] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleting the datastore file [datastore1] ba1daa0a-dcf8-4586-9964-a962fcc94196 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.046877] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e655449f-9761-4c25-820d-a32a2cd12fe4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.053348] env[63372]: DEBUG oslo_vmware.api [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 955.053348] env[63372]: value = "task-1024531" [ 955.053348] env[63372]: _type = "Task" [ 955.053348] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.060586] env[63372]: DEBUG oslo_vmware.api [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024531, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.248267] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024528, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.358165} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.248939] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.248939] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 955.248939] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 955.318608] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024529, 'name': Rename_Task, 'duration_secs': 0.146194} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.318937] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 955.319229] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ce0a0d8-d203-4a67-b23d-792030d5d0b4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.326230] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Waiting for the task: (returnval){ [ 955.326230] env[63372]: value = "task-1024533" [ 955.326230] env[63372]: _type = "Task" [ 955.326230] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.335691] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 955.335982] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024533, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.338859] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72d9a551-c5e6-4cad-8600-df2fb9e59016 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.344269] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 955.344269] env[63372]: value = "task-1024534" [ 955.344269] env[63372]: _type = "Task" [ 955.344269] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.354554] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024534, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.393919] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f57b98d-b518-47a5-a928-8285be576dcd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.402439] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b2add5-0b33-434d-8317-41e266f7d1b6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.435941] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b1be08-5c57-4ade-abd5-3d78e192ce03 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.445446] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015983b5-f952-4e59-8547-ff79ccd96975 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.462615] env[63372]: DEBUG nova.compute.provider_tree [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.566059] env[63372]: DEBUG oslo_vmware.api [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024531, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.386475} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.566523] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.566876] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 955.567214] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 955.567533] env[63372]: INFO nova.compute.manager [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Took 1.13 seconds to destroy the instance on the hypervisor. [ 955.567948] env[63372]: DEBUG oslo.service.loopingcall [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.568286] env[63372]: DEBUG nova.compute.manager [-] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 955.568452] env[63372]: DEBUG nova.network.neutron [-] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 955.807710] env[63372]: DEBUG nova.compute.manager [req-c7ac7bf6-34e2-4c84-9998-b019ab5fe6b3 req-d185c4e0-d1f2-4c15-8860-d31b69cfd95c service nova] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Received event network-vif-deleted-7d3f76e6-e43d-4af6-84df-522498d9968a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.808201] env[63372]: INFO nova.compute.manager [req-c7ac7bf6-34e2-4c84-9998-b019ab5fe6b3 req-d185c4e0-d1f2-4c15-8860-d31b69cfd95c service nova] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Neutron deleted interface 7d3f76e6-e43d-4af6-84df-522498d9968a; detaching it from the instance and deleting it from the info cache [ 955.808201] env[63372]: DEBUG nova.network.neutron [req-c7ac7bf6-34e2-4c84-9998-b019ab5fe6b3 req-d185c4e0-d1f2-4c15-8860-d31b69cfd95c service nova] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.835873] env[63372]: DEBUG oslo_vmware.api [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024533, 'name': PowerOnVM_Task, 'duration_secs': 0.496883} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.836180] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 955.836392] env[63372]: INFO nova.compute.manager [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Took 8.15 seconds to spawn the instance on the hypervisor. [ 955.836576] env[63372]: DEBUG nova.compute.manager [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.837360] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95bd6bc0-57bc-42a0-a536-c84394be4a50 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.853841] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024534, 'name': PowerOffVM_Task, 'duration_secs': 0.226319} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.854096] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 955.854286] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance '93a5d948-0629-4f53-a681-858d519acfa7' progress to 17 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 955.966013] env[63372]: DEBUG nova.scheduler.client.report [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.284020] env[63372]: DEBUG nova.virt.hardware [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 956.284315] env[63372]: DEBUG nova.virt.hardware [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 956.284484] env[63372]: DEBUG nova.virt.hardware [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.284670] env[63372]: DEBUG nova.virt.hardware [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 956.284818] env[63372]: DEBUG nova.virt.hardware [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.284969] env[63372]: DEBUG nova.virt.hardware [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 956.285193] env[63372]: DEBUG nova.virt.hardware [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 956.285355] env[63372]: DEBUG nova.virt.hardware [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 956.285547] env[63372]: DEBUG nova.virt.hardware [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 956.285733] env[63372]: DEBUG nova.virt.hardware [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 956.285915] env[63372]: DEBUG nova.virt.hardware [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.286778] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4d7ef4-9fa1-449c-9188-2f6d8a8b6ce5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.289366] env[63372]: DEBUG nova.network.neutron [-] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.296098] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dca043e-118f-4f83-a6ca-38e815b16545 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.311136] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:62:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d4ef133-b6f3-41d1-add4-92a1482195cf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '46264f2d-4aad-453f-b549-c48687e3b6ad', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 956.318876] env[63372]: DEBUG oslo.service.loopingcall [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.318876] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ee9c2d14-2c4a-42f2-89af-6ae2a71455fe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.320771] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 956.320997] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-add284c9-11ec-428a-91d6-92921f6283da {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.341160] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 956.341160] env[63372]: value = "task-1024535" [ 956.341160] env[63372]: _type = "Task" [ 956.341160] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.347965] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a81b8657-7be8-4ed1-968d-c298b2ca75c6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.364904] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 956.364904] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 956.364904] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.365549] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 956.365549] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.365549] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 956.365549] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 956.365735] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 956.365776] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 956.365952] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 956.366281] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.375419] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee4fe403-d8f6-4004-a25b-7272ed3d14b7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.391177] env[63372]: INFO nova.compute.manager [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Took 27.08 seconds to build instance. [ 956.392161] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024535, 'name': CreateVM_Task} progress is 15%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.396927] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 956.396927] env[63372]: value = "task-1024536" [ 956.396927] env[63372]: _type = "Task" [ 956.396927] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.404206] env[63372]: DEBUG nova.compute.manager [req-c7ac7bf6-34e2-4c84-9998-b019ab5fe6b3 req-d185c4e0-d1f2-4c15-8860-d31b69cfd95c service nova] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Detach interface failed, port_id=7d3f76e6-e43d-4af6-84df-522498d9968a, reason: Instance ba1daa0a-dcf8-4586-9964-a962fcc94196 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 956.413352] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.471214] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.279s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.471663] env[63372]: DEBUG nova.compute.manager [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 956.474778] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.491s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.476027] env[63372]: INFO nova.compute.claims [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 956.793072] env[63372]: INFO nova.compute.manager [-] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Took 1.22 seconds to deallocate network for instance. [ 956.852929] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024535, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.894204] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70d2caeb-b147-430b-88dd-a42f52ce299f tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Lock "c7c444cb-0aaa-4486-840a-e48fa255b421" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.602s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.913414] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024536, 'name': ReconfigVM_Task, 'duration_secs': 0.190732} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.913793] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance '93a5d948-0629-4f53-a681-858d519acfa7' progress to 33 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 956.980081] env[63372]: DEBUG nova.compute.utils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 956.983233] env[63372]: DEBUG nova.compute.manager [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 956.983421] env[63372]: DEBUG nova.network.neutron [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 957.032394] env[63372]: DEBUG nova.policy [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0cc3768e4e8c47a88565a21b7f05ad02', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '98581685387a4f1499ae6ed378af982c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 957.044615] env[63372]: DEBUG oslo_concurrency.lockutils [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Acquiring lock "c7c444cb-0aaa-4486-840a-e48fa255b421" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.044884] env[63372]: DEBUG oslo_concurrency.lockutils [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Lock "c7c444cb-0aaa-4486-840a-e48fa255b421" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.045110] env[63372]: DEBUG oslo_concurrency.lockutils [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Acquiring lock "c7c444cb-0aaa-4486-840a-e48fa255b421-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.045302] env[63372]: DEBUG oslo_concurrency.lockutils [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Lock "c7c444cb-0aaa-4486-840a-e48fa255b421-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.045473] env[63372]: DEBUG oslo_concurrency.lockutils [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Lock "c7c444cb-0aaa-4486-840a-e48fa255b421-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.047535] env[63372]: INFO nova.compute.manager [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Terminating instance [ 957.049205] env[63372]: DEBUG nova.compute.manager [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 957.049398] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 957.050310] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b0a8411-3698-4d11-b86a-df0acc1a92c5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.057814] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 957.058049] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-40cdf0d6-e29c-49d6-b283-e04ea2146c57 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.064535] env[63372]: DEBUG oslo_vmware.api [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Waiting for the task: (returnval){ [ 957.064535] env[63372]: value = "task-1024537" [ 957.064535] env[63372]: _type = "Task" [ 957.064535] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.072061] env[63372]: DEBUG oslo_vmware.api [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024537, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.299288] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.364069] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024535, 'name': CreateVM_Task, 'duration_secs': 0.554365} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.364069] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 957.364069] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.364069] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.364069] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 957.364069] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ae6adbd-65ea-40f8-aebc-6020274868c9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.364069] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 957.364069] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5228a5a7-45cd-4fba-b56e-f32fa9b33a58" [ 957.364069] env[63372]: _type = "Task" [ 957.364069] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.374145] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5228a5a7-45cd-4fba-b56e-f32fa9b33a58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.375144] env[63372]: DEBUG nova.network.neutron [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Successfully created port: 128a0686-7ca8-427f-9cf7-c63a2f370e1f {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 957.420179] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 957.420448] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 957.420607] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 957.420790] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 957.420936] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 957.421100] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 957.421306] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 957.421487] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 957.421698] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 957.421779] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 957.421952] env[63372]: DEBUG nova.virt.hardware [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 957.427967] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Reconfiguring VM instance instance-00000030 to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 957.428285] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c79eae95-2cff-4c0c-a980-7b807eab1f73 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.442041] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Volume attach. Driver type: vmdk {{(pid=63372) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 957.442267] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227427', 'volume_id': '3673857b-aabe-4a87-9f39-54d3076eee42', 'name': 'volume-3673857b-aabe-4a87-9f39-54d3076eee42', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4ee6c0ac-b7ac-49e7-b029-ca6e9a595912', 'attached_at': '', 'detached_at': '', 'volume_id': '3673857b-aabe-4a87-9f39-54d3076eee42', 'serial': '3673857b-aabe-4a87-9f39-54d3076eee42'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 957.443039] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20f4b5a0-48cf-4c96-a735-0ac9aa6e72e9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.459696] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b69088-4729-4d0c-85f8-f334de46092a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.462200] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 957.462200] env[63372]: value = "task-1024538" [ 957.462200] env[63372]: _type = "Task" [ 957.462200] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.486494] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] volume-3673857b-aabe-4a87-9f39-54d3076eee42/volume-3673857b-aabe-4a87-9f39-54d3076eee42.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.487503] env[63372]: DEBUG nova.compute.manager [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 957.492558] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0ca5881-c455-46a6-b831-96bbf3e0506e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.510463] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.515604] env[63372]: DEBUG oslo_vmware.api [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 957.515604] env[63372]: value = "task-1024539" [ 957.515604] env[63372]: _type = "Task" [ 957.515604] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.524264] env[63372]: DEBUG oslo_vmware.api [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024539, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.577367] env[63372]: DEBUG oslo_vmware.api [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024537, 'name': PowerOffVM_Task, 'duration_secs': 0.210879} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.577651] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 957.577951] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 957.578368] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a2053901-b6af-440c-a627-590058687c30 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.642027] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 957.642027] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 957.642027] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Deleting the datastore file [datastore2] c7c444cb-0aaa-4486-840a-e48fa255b421 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 957.642214] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-15c24f3b-9916-4a28-a8c3-004c2470279d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.649748] env[63372]: DEBUG oslo_vmware.api [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Waiting for the task: (returnval){ [ 957.649748] env[63372]: value = "task-1024541" [ 957.649748] env[63372]: _type = "Task" [ 957.649748] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.659912] env[63372]: DEBUG oslo_vmware.api [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024541, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.702289] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8735336-873e-4cb1-ac7f-536b6b2884fc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.709663] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9194117-6ebc-4661-be8c-c577bca94770 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.739274] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e5060d-33bf-45ae-9e1d-20cb0536ab39 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.746897] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13a53f2-e97d-453a-95fc-36982cf7988f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.760602] env[63372]: DEBUG nova.compute.provider_tree [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.875754] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5228a5a7-45cd-4fba-b56e-f32fa9b33a58, 'name': SearchDatastore_Task, 'duration_secs': 0.009533} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.876117] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.876299] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 957.876726] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.876885] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.877145] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 957.877436] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb62e2a5-cbfa-4943-a53b-a1b17c2b70d6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.888239] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 957.888424] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 957.889163] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff97f2f7-ac86-4355-bcf8-16b8d756a1f6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.894312] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 957.894312] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52c30547-af4c-9197-885f-4840c406dec9" [ 957.894312] env[63372]: _type = "Task" [ 957.894312] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.902196] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c30547-af4c-9197-885f-4840c406dec9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.972822] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024538, 'name': ReconfigVM_Task, 'duration_secs': 0.243635} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.973131] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Reconfigured VM instance instance-00000030 to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 957.973882] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9983e5-6646-49a1-9b81-7a005bb13965 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.996050] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 93a5d948-0629-4f53-a681-858d519acfa7/93a5d948-0629-4f53-a681-858d519acfa7.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.996337] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87fc2966-7cc2-4aa8-b7ee-691917187132 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.020921] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 958.020921] env[63372]: value = "task-1024542" [ 958.020921] env[63372]: _type = "Task" [ 958.020921] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.027211] env[63372]: DEBUG oslo_vmware.api [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024539, 'name': ReconfigVM_Task, 'duration_secs': 0.417201} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.027769] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Reconfigured VM instance instance-00000057 to attach disk [datastore2] volume-3673857b-aabe-4a87-9f39-54d3076eee42/volume-3673857b-aabe-4a87-9f39-54d3076eee42.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.035389] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69e13abc-b2a4-473b-88e2-9ebc8070a3b9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.046807] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024542, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.052886] env[63372]: DEBUG oslo_vmware.api [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 958.052886] env[63372]: value = "task-1024543" [ 958.052886] env[63372]: _type = "Task" [ 958.052886] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.061554] env[63372]: DEBUG oslo_vmware.api [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024543, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.159784] env[63372]: DEBUG oslo_vmware.api [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Task: {'id': task-1024541, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187059} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.160097] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 958.160306] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 958.160490] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 958.160662] env[63372]: INFO nova.compute.manager [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Took 1.11 seconds to destroy the instance on the hypervisor. [ 958.160929] env[63372]: DEBUG oslo.service.loopingcall [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.161189] env[63372]: DEBUG nova.compute.manager [-] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 958.161277] env[63372]: DEBUG nova.network.neutron [-] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 958.263225] env[63372]: DEBUG nova.scheduler.client.report [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.404714] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c30547-af4c-9197-885f-4840c406dec9, 'name': SearchDatastore_Task, 'duration_secs': 0.013401} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.405808] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be8338b5-9d3c-4f8b-856f-85190bb351eb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.411993] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 958.411993] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fbcbb2-af90-7475-a497-6abf74ff0e6e" [ 958.411993] env[63372]: _type = "Task" [ 958.411993] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.420380] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fbcbb2-af90-7475-a497-6abf74ff0e6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.456163] env[63372]: DEBUG nova.compute.manager [req-14abf468-b471-43df-ad31-b53f6d70967f req-eb78b2fd-5d85-474b-8965-793624b30eb2 service nova] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Received event network-vif-deleted-eeee4694-73df-47f5-bd8c-82b44acebe2f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.456396] env[63372]: INFO nova.compute.manager [req-14abf468-b471-43df-ad31-b53f6d70967f req-eb78b2fd-5d85-474b-8965-793624b30eb2 service nova] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Neutron deleted interface eeee4694-73df-47f5-bd8c-82b44acebe2f; detaching it from the instance and deleting it from the info cache [ 958.456575] env[63372]: DEBUG nova.network.neutron [req-14abf468-b471-43df-ad31-b53f6d70967f req-eb78b2fd-5d85-474b-8965-793624b30eb2 service nova] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.518021] env[63372]: DEBUG nova.compute.manager [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 958.534212] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024542, 'name': ReconfigVM_Task, 'duration_secs': 0.413364} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.534510] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 93a5d948-0629-4f53-a681-858d519acfa7/93a5d948-0629-4f53-a681-858d519acfa7.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.534750] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance '93a5d948-0629-4f53-a681-858d519acfa7' progress to 50 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 958.546138] env[63372]: DEBUG nova.virt.hardware [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 958.546389] env[63372]: DEBUG nova.virt.hardware [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 958.547036] env[63372]: DEBUG nova.virt.hardware [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 958.547036] env[63372]: DEBUG nova.virt.hardware [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 958.547036] env[63372]: DEBUG nova.virt.hardware [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 958.547036] env[63372]: DEBUG nova.virt.hardware [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 958.547249] env[63372]: DEBUG nova.virt.hardware [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 958.547383] env[63372]: DEBUG nova.virt.hardware [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 958.547551] env[63372]: DEBUG nova.virt.hardware [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 958.547713] env[63372]: DEBUG nova.virt.hardware [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 958.547925] env[63372]: DEBUG nova.virt.hardware [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 958.548779] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e05d87-579b-4f8c-8923-16bbebbef145 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.559474] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e58dcd0a-abde-467d-a470-7c8353d1044c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.576041] env[63372]: DEBUG oslo_vmware.api [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024543, 'name': ReconfigVM_Task, 'duration_secs': 0.140663} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.576560] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227427', 'volume_id': '3673857b-aabe-4a87-9f39-54d3076eee42', 'name': 'volume-3673857b-aabe-4a87-9f39-54d3076eee42', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4ee6c0ac-b7ac-49e7-b029-ca6e9a595912', 'attached_at': '', 'detached_at': '', 'volume_id': '3673857b-aabe-4a87-9f39-54d3076eee42', 'serial': '3673857b-aabe-4a87-9f39-54d3076eee42'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 958.767832] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.293s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.768408] env[63372]: DEBUG nova.compute.manager [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 958.771092] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.628s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.771322] env[63372]: DEBUG nova.objects.instance [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Lazy-loading 'resources' on Instance uuid 662d9fdf-ec0d-4466-8d15-abc942e4dad8 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.922810] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fbcbb2-af90-7475-a497-6abf74ff0e6e, 'name': SearchDatastore_Task, 'duration_secs': 0.013291} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.923110] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.923371] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 61f7a052-9a42-4df3-a568-eab197fc5c4f/61f7a052-9a42-4df3-a568-eab197fc5c4f.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 958.923627] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3940c911-5a06-43f7-abad-943b750ba683 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.926872] env[63372]: DEBUG nova.network.neutron [-] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.930481] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 958.930481] env[63372]: value = "task-1024544" [ 958.930481] env[63372]: _type = "Task" [ 958.930481] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.940492] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024544, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.962042] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64e28b31-1307-44e8-8bf9-d4391eeea667 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.969647] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0418c179-06af-48c9-b692-4e3e20cee5b1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.998851] env[63372]: DEBUG nova.compute.manager [req-14abf468-b471-43df-ad31-b53f6d70967f req-eb78b2fd-5d85-474b-8965-793624b30eb2 service nova] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Detach interface failed, port_id=eeee4694-73df-47f5-bd8c-82b44acebe2f, reason: Instance c7c444cb-0aaa-4486-840a-e48fa255b421 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 959.007453] env[63372]: DEBUG nova.network.neutron [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Successfully updated port: 128a0686-7ca8-427f-9cf7-c63a2f370e1f {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 959.044631] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b72f61-c7f6-4fe5-84a8-5718c3a5b187 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.065295] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3693c2e9-14e5-4d75-8446-5a1cfe47ac9b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.088170] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance '93a5d948-0629-4f53-a681-858d519acfa7' progress to 67 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 959.274486] env[63372]: DEBUG nova.compute.utils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 959.279952] env[63372]: DEBUG nova.compute.manager [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 959.279952] env[63372]: DEBUG nova.network.neutron [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 959.320476] env[63372]: DEBUG nova.policy [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9592bafdf42843c18cb71d97f9fbd60a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e517915d5e7a4e3e924cc95c55170a66', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 959.429746] env[63372]: INFO nova.compute.manager [-] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Took 1.27 seconds to deallocate network for instance. [ 959.445009] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024544, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461704} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.447356] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 61f7a052-9a42-4df3-a568-eab197fc5c4f/61f7a052-9a42-4df3-a568-eab197fc5c4f.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 959.447578] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 959.448029] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dfe1e148-3807-45e7-9507-26d090900b42 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.454869] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 959.454869] env[63372]: value = "task-1024545" [ 959.454869] env[63372]: _type = "Task" [ 959.454869] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.465384] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024545, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.476145] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-225616e3-dfc3-4c9a-aa55-4064b68f11d8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.483762] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88860eb2-dd9b-4ef6-858e-fb20959619f6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.513874] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.514099] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.514180] env[63372]: DEBUG nova.network.neutron [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 959.515914] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d63fb1e-add1-4614-94b9-44da922eaf49 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.523978] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbad161a-c649-4f7f-adf6-71babefab5ae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.539544] env[63372]: DEBUG nova.compute.provider_tree [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.627802] env[63372]: DEBUG nova.objects.instance [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lazy-loading 'flavor' on Instance uuid 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.647075] env[63372]: DEBUG nova.network.neutron [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Port 8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f binding to destination host cpu-1 is already ACTIVE {{(pid=63372) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 959.689967] env[63372]: DEBUG nova.network.neutron [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Successfully created port: eef17039-cc41-41f1-9c94-291f90803267 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 959.780661] env[63372]: DEBUG nova.compute.manager [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 959.942156] env[63372]: DEBUG oslo_concurrency.lockutils [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.965120] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024545, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086931} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.965395] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 959.966155] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0501ea63-9dc8-40e3-acdb-e670f2f71754 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.988088] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 61f7a052-9a42-4df3-a568-eab197fc5c4f/61f7a052-9a42-4df3-a568-eab197fc5c4f.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.988362] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df32809e-5f8d-426f-8c9e-4f2c9f4c2c59 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.008150] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 960.008150] env[63372]: value = "task-1024546" [ 960.008150] env[63372]: _type = "Task" [ 960.008150] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.015914] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024546, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.042019] env[63372]: DEBUG nova.scheduler.client.report [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.091512] env[63372]: DEBUG nova.network.neutron [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 960.132872] env[63372]: DEBUG oslo_concurrency.lockutils [None req-43f5dff0-de55-43f6-b612-01d897e961bb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.312s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.233272] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.233272] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.233272] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.233448] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.233585] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.243601] env[63372]: INFO nova.compute.manager [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Terminating instance [ 960.250016] env[63372]: DEBUG nova.compute.manager [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 960.250016] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 960.250016] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d2eb128-0e4d-4ef6-8d21-ca4b64baa02d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.257330] env[63372]: DEBUG oslo_vmware.api [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 960.257330] env[63372]: value = "task-1024547" [ 960.257330] env[63372]: _type = "Task" [ 960.257330] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.267966] env[63372]: DEBUG oslo_vmware.api [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024547, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.484698] env[63372]: DEBUG nova.compute.manager [req-3ea2b630-b30e-42fb-a1d2-37f5ce52671c req-1607949e-3e4c-46cd-a569-182b83321c0d service nova] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Received event network-vif-plugged-128a0686-7ca8-427f-9cf7-c63a2f370e1f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.484951] env[63372]: DEBUG oslo_concurrency.lockutils [req-3ea2b630-b30e-42fb-a1d2-37f5ce52671c req-1607949e-3e4c-46cd-a569-182b83321c0d service nova] Acquiring lock "b7732621-7e58-40a3-b723-5c66df6f74a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.485181] env[63372]: DEBUG oslo_concurrency.lockutils [req-3ea2b630-b30e-42fb-a1d2-37f5ce52671c req-1607949e-3e4c-46cd-a569-182b83321c0d service nova] Lock "b7732621-7e58-40a3-b723-5c66df6f74a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.485354] env[63372]: DEBUG oslo_concurrency.lockutils [req-3ea2b630-b30e-42fb-a1d2-37f5ce52671c req-1607949e-3e4c-46cd-a569-182b83321c0d service nova] Lock "b7732621-7e58-40a3-b723-5c66df6f74a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.485522] env[63372]: DEBUG nova.compute.manager [req-3ea2b630-b30e-42fb-a1d2-37f5ce52671c req-1607949e-3e4c-46cd-a569-182b83321c0d service nova] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] No waiting events found dispatching network-vif-plugged-128a0686-7ca8-427f-9cf7-c63a2f370e1f {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 960.485798] env[63372]: WARNING nova.compute.manager [req-3ea2b630-b30e-42fb-a1d2-37f5ce52671c req-1607949e-3e4c-46cd-a569-182b83321c0d service nova] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Received unexpected event network-vif-plugged-128a0686-7ca8-427f-9cf7-c63a2f370e1f for instance with vm_state building and task_state spawning. [ 960.485857] env[63372]: DEBUG nova.compute.manager [req-3ea2b630-b30e-42fb-a1d2-37f5ce52671c req-1607949e-3e4c-46cd-a569-182b83321c0d service nova] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Received event network-changed-128a0686-7ca8-427f-9cf7-c63a2f370e1f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.486119] env[63372]: DEBUG nova.compute.manager [req-3ea2b630-b30e-42fb-a1d2-37f5ce52671c req-1607949e-3e4c-46cd-a569-182b83321c0d service nova] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Refreshing instance network info cache due to event network-changed-128a0686-7ca8-427f-9cf7-c63a2f370e1f. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 960.486168] env[63372]: DEBUG oslo_concurrency.lockutils [req-3ea2b630-b30e-42fb-a1d2-37f5ce52671c req-1607949e-3e4c-46cd-a569-182b83321c0d service nova] Acquiring lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.487122] env[63372]: DEBUG nova.network.neutron [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance_info_cache with network_info: [{"id": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "address": "fa:16:3e:28:02:b8", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap128a0686-7c", "ovs_interfaceid": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.518097] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024546, 'name': ReconfigVM_Task, 'duration_secs': 0.288955} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.518510] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 61f7a052-9a42-4df3-a568-eab197fc5c4f/61f7a052-9a42-4df3-a568-eab197fc5c4f.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 960.519137] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5504e555-03cf-4b15-b4c4-12c064d518b6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.525092] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 960.525092] env[63372]: value = "task-1024548" [ 960.525092] env[63372]: _type = "Task" [ 960.525092] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.532801] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024548, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.546968] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.776s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.549435] env[63372]: DEBUG oslo_concurrency.lockutils [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.538s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.549747] env[63372]: DEBUG nova.objects.instance [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Lazy-loading 'resources' on Instance uuid ed2cefa1-3604-43c5-9860-0fb49133bbb4 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.572090] env[63372]: INFO nova.scheduler.client.report [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Deleted allocations for instance 662d9fdf-ec0d-4466-8d15-abc942e4dad8 [ 960.669774] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "93a5d948-0629-4f53-a681-858d519acfa7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.670060] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "93a5d948-0629-4f53-a681-858d519acfa7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.670273] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "93a5d948-0629-4f53-a681-858d519acfa7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.766175] env[63372]: DEBUG oslo_vmware.api [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024547, 'name': PowerOffVM_Task, 'duration_secs': 0.200641} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.766479] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 960.766656] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Volume detach. Driver type: vmdk {{(pid=63372) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 960.766853] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227427', 'volume_id': '3673857b-aabe-4a87-9f39-54d3076eee42', 'name': 'volume-3673857b-aabe-4a87-9f39-54d3076eee42', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4ee6c0ac-b7ac-49e7-b029-ca6e9a595912', 'attached_at': '', 'detached_at': '', 'volume_id': '3673857b-aabe-4a87-9f39-54d3076eee42', 'serial': '3673857b-aabe-4a87-9f39-54d3076eee42'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 960.767624] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a630d8-7065-4137-bf76-f406a2218e3e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.788787] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41b0c4a-cb9d-48a1-9b46-7c32b8a65bd5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.792394] env[63372]: DEBUG nova.compute.manager [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 960.798718] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad105660-f590-4020-9122-df0287840ead {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.822450] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c8929c-adcf-47f0-ae91-226f091d1c28 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.837312] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] The volume has not been displaced from its original location: [datastore2] volume-3673857b-aabe-4a87-9f39-54d3076eee42/volume-3673857b-aabe-4a87-9f39-54d3076eee42.vmdk. No consolidation needed. {{(pid=63372) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 960.842629] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Reconfiguring VM instance instance-00000057 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 960.844871] env[63372]: DEBUG nova.virt.hardware [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 960.845111] env[63372]: DEBUG nova.virt.hardware [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 960.845276] env[63372]: DEBUG nova.virt.hardware [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 960.845462] env[63372]: DEBUG nova.virt.hardware [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 960.845609] env[63372]: DEBUG nova.virt.hardware [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 960.845756] env[63372]: DEBUG nova.virt.hardware [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 960.845963] env[63372]: DEBUG nova.virt.hardware [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 960.846186] env[63372]: DEBUG nova.virt.hardware [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 960.846377] env[63372]: DEBUG nova.virt.hardware [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 960.846545] env[63372]: DEBUG nova.virt.hardware [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 960.846719] env[63372]: DEBUG nova.virt.hardware [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 960.846971] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c01ade5f-9fa9-47cd-b3fb-6341a910d332 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.860210] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5bf6660-18dc-4a95-8f87-84f0d2af65d4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.867721] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7bbc38d-c55f-4138-96bc-99ff18760f1d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.871845] env[63372]: DEBUG oslo_vmware.api [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 960.871845] env[63372]: value = "task-1024549" [ 960.871845] env[63372]: _type = "Task" [ 960.871845] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.886915] env[63372]: DEBUG oslo_vmware.api [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024549, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.990007] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.990504] env[63372]: DEBUG nova.compute.manager [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Instance network_info: |[{"id": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "address": "fa:16:3e:28:02:b8", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap128a0686-7c", "ovs_interfaceid": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 960.990994] env[63372]: DEBUG oslo_concurrency.lockutils [req-3ea2b630-b30e-42fb-a1d2-37f5ce52671c req-1607949e-3e4c-46cd-a569-182b83321c0d service nova] Acquired lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.991332] env[63372]: DEBUG nova.network.neutron [req-3ea2b630-b30e-42fb-a1d2-37f5ce52671c req-1607949e-3e4c-46cd-a569-182b83321c0d service nova] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Refreshing network info cache for port 128a0686-7ca8-427f-9cf7-c63a2f370e1f {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 960.992836] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:02:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '128a0686-7ca8-427f-9cf7-c63a2f370e1f', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.000580] env[63372]: DEBUG oslo.service.loopingcall [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.001616] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 961.001846] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-92f0431e-34cb-47de-ac45-a1c967c5348a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.021984] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.021984] env[63372]: value = "task-1024550" [ 961.021984] env[63372]: _type = "Task" [ 961.021984] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.031510] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024550, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.036529] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024548, 'name': Rename_Task, 'duration_secs': 0.136831} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.036807] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 961.037074] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-191b688e-6dd5-491c-8af1-2e728baddfc4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.045384] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 961.045384] env[63372]: value = "task-1024551" [ 961.045384] env[63372]: _type = "Task" [ 961.045384] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.054201] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024551, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.080296] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c68ddc75-73a7-4586-bab1-561579fedf37 tempest-ServerShowV254Test-963585099 tempest-ServerShowV254Test-963585099-project-member] Lock "662d9fdf-ec0d-4466-8d15-abc942e4dad8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.648s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.225601] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc577e3-aa2d-45bb-8fe0-9e8454ea4c89 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.233912] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03849325-e92e-4a52-8697-272c690a7058 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.270173] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9eb569-e05a-4610-9186-f5dd1ed2b90b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.278916] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bbb0f3c-2bf2-4a68-9e3d-bae87f51615f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.296603] env[63372]: DEBUG nova.compute.provider_tree [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.382180] env[63372]: DEBUG oslo_vmware.api [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024549, 'name': ReconfigVM_Task, 'duration_secs': 0.21794} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.382484] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Reconfigured VM instance instance-00000057 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 961.387154] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc29724f-f508-484f-a4ee-84a1eaa5dcdb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.408606] env[63372]: DEBUG oslo_vmware.api [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 961.408606] env[63372]: value = "task-1024552" [ 961.408606] env[63372]: _type = "Task" [ 961.408606] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.418572] env[63372]: DEBUG oslo_vmware.api [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024552, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.532278] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024550, 'name': CreateVM_Task, 'duration_secs': 0.493441} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.532660] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 961.533208] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.533386] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.533711] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 961.533971] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf590f5a-d23b-4e2c-b29d-30215b7eec08 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.539410] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 961.539410] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52c39bfd-24f3-d9d9-21b8-d1f4f78c09d0" [ 961.539410] env[63372]: _type = "Task" [ 961.539410] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.547657] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c39bfd-24f3-d9d9-21b8-d1f4f78c09d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.555378] env[63372]: DEBUG oslo_vmware.api [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024551, 'name': PowerOnVM_Task, 'duration_secs': 0.508925} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.555671] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 961.555885] env[63372]: DEBUG nova.compute.manager [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 961.556712] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719ca6a9-30ca-45db-b377-4c4bf5d56cf8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.703504] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.703760] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.704055] env[63372]: DEBUG nova.network.neutron [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 961.799071] env[63372]: DEBUG nova.scheduler.client.report [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 961.921014] env[63372]: DEBUG oslo_vmware.api [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024552, 'name': ReconfigVM_Task, 'duration_secs': 0.198268} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.921356] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227427', 'volume_id': '3673857b-aabe-4a87-9f39-54d3076eee42', 'name': 'volume-3673857b-aabe-4a87-9f39-54d3076eee42', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4ee6c0ac-b7ac-49e7-b029-ca6e9a595912', 'attached_at': '', 'detached_at': '', 'volume_id': '3673857b-aabe-4a87-9f39-54d3076eee42', 'serial': '3673857b-aabe-4a87-9f39-54d3076eee42'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 961.921644] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 961.922533] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998423b2-8e2a-4554-b7b8-04b2a2d7af4e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.929459] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 961.929702] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-90a050e3-5db4-4055-aa11-d2882e94fcdf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.990157] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 961.990519] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 961.990646] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleting the datastore file [datastore2] 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 961.990846] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f48e667b-47d3-4a56-ab40-8d95ab833df0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.997589] env[63372]: DEBUG oslo_vmware.api [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 961.997589] env[63372]: value = "task-1024554" [ 961.997589] env[63372]: _type = "Task" [ 961.997589] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.006061] env[63372]: DEBUG oslo_vmware.api [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024554, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.031016] env[63372]: DEBUG nova.network.neutron [req-3ea2b630-b30e-42fb-a1d2-37f5ce52671c req-1607949e-3e4c-46cd-a569-182b83321c0d service nova] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updated VIF entry in instance network info cache for port 128a0686-7ca8-427f-9cf7-c63a2f370e1f. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 962.031432] env[63372]: DEBUG nova.network.neutron [req-3ea2b630-b30e-42fb-a1d2-37f5ce52671c req-1607949e-3e4c-46cd-a569-182b83321c0d service nova] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance_info_cache with network_info: [{"id": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "address": "fa:16:3e:28:02:b8", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap128a0686-7c", "ovs_interfaceid": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.050735] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c39bfd-24f3-d9d9-21b8-d1f4f78c09d0, 'name': SearchDatastore_Task, 'duration_secs': 0.010732} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.051057] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.051305] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 962.051539] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.051914] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.051914] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.052129] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1116680-e2bb-48e7-ba8f-d32b306678b3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.059600] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.059780] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 962.060501] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5a07348-b769-462f-b666-a75c64417522 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.069414] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 962.069414] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5219a68c-32cc-e679-6045-be3d93f166f5" [ 962.069414] env[63372]: _type = "Task" [ 962.069414] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.074124] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.079360] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5219a68c-32cc-e679-6045-be3d93f166f5, 'name': SearchDatastore_Task, 'duration_secs': 0.00844} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.080096] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c14ccd4-3bd2-4d80-b3aa-1dcff42371ff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.085046] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 962.085046] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523bd813-a0c9-471d-f7df-29cc2f67b837" [ 962.085046] env[63372]: _type = "Task" [ 962.085046] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.091917] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523bd813-a0c9-471d-f7df-29cc2f67b837, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.113386] env[63372]: DEBUG nova.network.neutron [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Successfully updated port: eef17039-cc41-41f1-9c94-291f90803267 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 962.303361] env[63372]: DEBUG oslo_concurrency.lockutils [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.754s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.306055] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.810s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.307035] env[63372]: INFO nova.compute.claims [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 962.332020] env[63372]: INFO nova.scheduler.client.report [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Deleted allocations for instance ed2cefa1-3604-43c5-9860-0fb49133bbb4 [ 962.506871] env[63372]: DEBUG oslo_vmware.api [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024554, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141502} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.507145] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 962.507359] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 962.507527] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 962.507702] env[63372]: INFO nova.compute.manager [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Took 2.26 seconds to destroy the instance on the hypervisor. [ 962.508047] env[63372]: DEBUG oslo.service.loopingcall [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.508263] env[63372]: DEBUG nova.compute.manager [-] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 962.508359] env[63372]: DEBUG nova.network.neutron [-] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 962.510678] env[63372]: DEBUG nova.network.neutron [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance_info_cache with network_info: [{"id": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "address": "fa:16:3e:3f:99:77", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c4c74ba-ad", "ovs_interfaceid": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.514563] env[63372]: DEBUG nova.compute.manager [req-ebb5d029-bfc0-4710-bc36-33d6f17bd2fa req-8f1308af-949d-4d24-96bf-ad37e7ea4ffb service nova] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Received event network-vif-plugged-eef17039-cc41-41f1-9c94-291f90803267 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.514563] env[63372]: DEBUG oslo_concurrency.lockutils [req-ebb5d029-bfc0-4710-bc36-33d6f17bd2fa req-8f1308af-949d-4d24-96bf-ad37e7ea4ffb service nova] Acquiring lock "6beadbab-2cc1-4b69-95c9-e1a0ea11045f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.514563] env[63372]: DEBUG oslo_concurrency.lockutils [req-ebb5d029-bfc0-4710-bc36-33d6f17bd2fa req-8f1308af-949d-4d24-96bf-ad37e7ea4ffb service nova] Lock "6beadbab-2cc1-4b69-95c9-e1a0ea11045f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.514563] env[63372]: DEBUG oslo_concurrency.lockutils [req-ebb5d029-bfc0-4710-bc36-33d6f17bd2fa req-8f1308af-949d-4d24-96bf-ad37e7ea4ffb service nova] Lock "6beadbab-2cc1-4b69-95c9-e1a0ea11045f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.514563] env[63372]: DEBUG nova.compute.manager [req-ebb5d029-bfc0-4710-bc36-33d6f17bd2fa req-8f1308af-949d-4d24-96bf-ad37e7ea4ffb service nova] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] No waiting events found dispatching network-vif-plugged-eef17039-cc41-41f1-9c94-291f90803267 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 962.514563] env[63372]: WARNING nova.compute.manager [req-ebb5d029-bfc0-4710-bc36-33d6f17bd2fa req-8f1308af-949d-4d24-96bf-ad37e7ea4ffb service nova] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Received unexpected event network-vif-plugged-eef17039-cc41-41f1-9c94-291f90803267 for instance with vm_state building and task_state spawning. [ 962.514563] env[63372]: DEBUG nova.compute.manager [req-ebb5d029-bfc0-4710-bc36-33d6f17bd2fa req-8f1308af-949d-4d24-96bf-ad37e7ea4ffb service nova] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Received event network-changed-eef17039-cc41-41f1-9c94-291f90803267 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.514979] env[63372]: DEBUG nova.compute.manager [req-ebb5d029-bfc0-4710-bc36-33d6f17bd2fa req-8f1308af-949d-4d24-96bf-ad37e7ea4ffb service nova] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Refreshing instance network info cache due to event network-changed-eef17039-cc41-41f1-9c94-291f90803267. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 962.514979] env[63372]: DEBUG oslo_concurrency.lockutils [req-ebb5d029-bfc0-4710-bc36-33d6f17bd2fa req-8f1308af-949d-4d24-96bf-ad37e7ea4ffb service nova] Acquiring lock "refresh_cache-6beadbab-2cc1-4b69-95c9-e1a0ea11045f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.515122] env[63372]: DEBUG oslo_concurrency.lockutils [req-ebb5d029-bfc0-4710-bc36-33d6f17bd2fa req-8f1308af-949d-4d24-96bf-ad37e7ea4ffb service nova] Acquired lock "refresh_cache-6beadbab-2cc1-4b69-95c9-e1a0ea11045f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.515183] env[63372]: DEBUG nova.network.neutron [req-ebb5d029-bfc0-4710-bc36-33d6f17bd2fa req-8f1308af-949d-4d24-96bf-ad37e7ea4ffb service nova] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Refreshing network info cache for port eef17039-cc41-41f1-9c94-291f90803267 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 962.533847] env[63372]: DEBUG oslo_concurrency.lockutils [req-3ea2b630-b30e-42fb-a1d2-37f5ce52671c req-1607949e-3e4c-46cd-a569-182b83321c0d service nova] Releasing lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.595941] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523bd813-a0c9-471d-f7df-29cc2f67b837, 'name': SearchDatastore_Task, 'duration_secs': 0.009716} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.596225] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.596494] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] b7732621-7e58-40a3-b723-5c66df6f74a1/b7732621-7e58-40a3-b723-5c66df6f74a1.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 962.596758] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-839d06e0-5756-4204-9a44-7b21be922e56 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.603231] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 962.603231] env[63372]: value = "task-1024555" [ 962.603231] env[63372]: _type = "Task" [ 962.603231] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.610890] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024555, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.616068] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "refresh_cache-6beadbab-2cc1-4b69-95c9-e1a0ea11045f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.842113] env[63372]: DEBUG oslo_concurrency.lockutils [None req-793d14c6-9f70-40b5-bb55-a9d0eaff4b7a tempest-ServerTagsTestJSON-468648300 tempest-ServerTagsTestJSON-468648300-project-member] Lock "ed2cefa1-3604-43c5-9860-0fb49133bbb4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.217s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.017716] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.108810] env[63372]: DEBUG nova.network.neutron [req-ebb5d029-bfc0-4710-bc36-33d6f17bd2fa req-8f1308af-949d-4d24-96bf-ad37e7ea4ffb service nova] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 963.116990] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024555, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487456} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.117262] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] b7732621-7e58-40a3-b723-5c66df6f74a1/b7732621-7e58-40a3-b723-5c66df6f74a1.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 963.117487] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 963.117759] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f26d87b4-b046-4118-a8f4-250f7d394c04 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.124424] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 963.124424] env[63372]: value = "task-1024556" [ 963.124424] env[63372]: _type = "Task" [ 963.124424] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.132370] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024556, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.217141] env[63372]: DEBUG nova.network.neutron [-] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.314897] env[63372]: DEBUG nova.network.neutron [req-ebb5d029-bfc0-4710-bc36-33d6f17bd2fa req-8f1308af-949d-4d24-96bf-ad37e7ea4ffb service nova] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.499181] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef17898e-1d2f-464c-8495-76abc65cd893 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.506663] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cdccc0d-f774-4e4d-8de6-b19de78c587c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.543087] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e4de75-932a-4fbf-8aab-25b825845440 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.551203] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0154f7-df44-4624-8b86-6063dc6b9cd0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.557094] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96c2438-a16b-49a8-9ac5-fb195fa5f7e4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.584027] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de877460-d27e-4554-bcd9-3d1dbf116fc0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.587026] env[63372]: DEBUG nova.compute.provider_tree [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.593634] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance '93a5d948-0629-4f53-a681-858d519acfa7' progress to 83 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 963.634816] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024556, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078165} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.635131] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 963.635916] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecfe5fd6-f4b2-4b45-9472-07f9d7f2dd19 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.658246] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] b7732621-7e58-40a3-b723-5c66df6f74a1/b7732621-7e58-40a3-b723-5c66df6f74a1.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.659283] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a81b77f4-201d-4c90-adb3-2bfa5cf1b726 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.678714] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 963.678714] env[63372]: value = "task-1024557" [ 963.678714] env[63372]: _type = "Task" [ 963.678714] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.686859] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024557, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.721178] env[63372]: INFO nova.compute.manager [-] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Took 1.21 seconds to deallocate network for instance. [ 963.791661] env[63372]: DEBUG oslo_concurrency.lockutils [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "61f7a052-9a42-4df3-a568-eab197fc5c4f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.792041] env[63372]: DEBUG oslo_concurrency.lockutils [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "61f7a052-9a42-4df3-a568-eab197fc5c4f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.792312] env[63372]: DEBUG oslo_concurrency.lockutils [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "61f7a052-9a42-4df3-a568-eab197fc5c4f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.792554] env[63372]: DEBUG oslo_concurrency.lockutils [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "61f7a052-9a42-4df3-a568-eab197fc5c4f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.792752] env[63372]: DEBUG oslo_concurrency.lockutils [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "61f7a052-9a42-4df3-a568-eab197fc5c4f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.795194] env[63372]: INFO nova.compute.manager [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Terminating instance [ 963.797348] env[63372]: DEBUG nova.compute.manager [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 963.797568] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 963.798471] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0000c06d-7269-4602-af4d-5cbc50bf30e7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.806833] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 963.807092] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49a46adb-0e64-4b5d-8128-c358032d0128 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.813463] env[63372]: DEBUG oslo_vmware.api [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 963.813463] env[63372]: value = "task-1024558" [ 963.813463] env[63372]: _type = "Task" [ 963.813463] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.822626] env[63372]: DEBUG oslo_concurrency.lockutils [req-ebb5d029-bfc0-4710-bc36-33d6f17bd2fa req-8f1308af-949d-4d24-96bf-ad37e7ea4ffb service nova] Releasing lock "refresh_cache-6beadbab-2cc1-4b69-95c9-e1a0ea11045f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.823051] env[63372]: DEBUG oslo_vmware.api [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024558, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.823336] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquired lock "refresh_cache-6beadbab-2cc1-4b69-95c9-e1a0ea11045f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.823506] env[63372]: DEBUG nova.network.neutron [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 964.090375] env[63372]: DEBUG nova.scheduler.client.report [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.099810] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 964.100087] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8bf6eee-9b75-4712-a1cf-32f9a81d18ed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.106703] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 964.106703] env[63372]: value = "task-1024559" [ 964.106703] env[63372]: _type = "Task" [ 964.106703] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.118344] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024559, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.191514] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024557, 'name': ReconfigVM_Task, 'duration_secs': 0.294923} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.191815] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Reconfigured VM instance instance-0000005a to attach disk [datastore2] b7732621-7e58-40a3-b723-5c66df6f74a1/b7732621-7e58-40a3-b723-5c66df6f74a1.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.192483] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-94eb229e-0e59-449c-8cd6-34db10f1f04c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.199544] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 964.199544] env[63372]: value = "task-1024560" [ 964.199544] env[63372]: _type = "Task" [ 964.199544] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.207611] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024560, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.267671] env[63372]: INFO nova.compute.manager [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Took 0.54 seconds to detach 1 volumes for instance. [ 964.326063] env[63372]: DEBUG oslo_vmware.api [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024558, 'name': PowerOffVM_Task, 'duration_secs': 0.358025} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.326063] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 964.326063] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 964.326063] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-adf47012-66da-4fe8-b682-c4bad4548fdf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.404645] env[63372]: DEBUG nova.network.neutron [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 964.410585] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 964.410936] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 964.411257] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleting the datastore file [datastore2] 61f7a052-9a42-4df3-a568-eab197fc5c4f {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.411611] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79235b94-3bf7-4a87-8017-367b0b2b2906 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.417191] env[63372]: DEBUG oslo_vmware.api [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 964.417191] env[63372]: value = "task-1024562" [ 964.417191] env[63372]: _type = "Task" [ 964.417191] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.426480] env[63372]: DEBUG oslo_vmware.api [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024562, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.454223] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.454681] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.596949] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.291s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.597458] env[63372]: DEBUG nova.compute.manager [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 964.604070] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.302s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.604070] env[63372]: DEBUG nova.objects.instance [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lazy-loading 'resources' on Instance uuid ba1daa0a-dcf8-4586-9964-a962fcc94196 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.630022] env[63372]: DEBUG oslo_vmware.api [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024559, 'name': PowerOnVM_Task, 'duration_secs': 0.506777} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.630022] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 964.630022] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf55ff56-f753-47bd-8fd8-2ab3a6573f70 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance '93a5d948-0629-4f53-a681-858d519acfa7' progress to 100 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 964.710850] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024560, 'name': Rename_Task, 'duration_secs': 0.208764} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.711182] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 964.711436] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce56b22c-e592-49fc-853c-880ccdae78c1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.718114] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 964.718114] env[63372]: value = "task-1024563" [ 964.718114] env[63372]: _type = "Task" [ 964.718114] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.726301] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024563, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.776676] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.894119] env[63372]: DEBUG nova.compute.manager [req-99aebc2b-ba40-4607-be83-321befc8d74a req-d3ff8915-0682-4db4-ac9d-631d3d1afa42 service nova] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Received event network-vif-deleted-a6afe51b-6c5f-4f54-9657-97c43e85cdba {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.927421] env[63372]: DEBUG oslo_vmware.api [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024562, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.337706} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.927699] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 964.927886] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 964.928104] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 964.928286] env[63372]: INFO nova.compute.manager [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 964.928525] env[63372]: DEBUG oslo.service.loopingcall [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 964.928718] env[63372]: DEBUG nova.compute.manager [-] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 964.928818] env[63372]: DEBUG nova.network.neutron [-] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 964.959436] env[63372]: DEBUG nova.compute.manager [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 964.978901] env[63372]: DEBUG nova.network.neutron [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Updating instance_info_cache with network_info: [{"id": "eef17039-cc41-41f1-9c94-291f90803267", "address": "fa:16:3e:87:d7:99", "network": {"id": "cbdcb945-f0a5-49b1-9b18-fcd8e4ca9728", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1068757189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e517915d5e7a4e3e924cc95c55170a66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeef17039-cc", "ovs_interfaceid": "eef17039-cc41-41f1-9c94-291f90803267", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.109513] env[63372]: DEBUG nova.compute.utils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 965.112802] env[63372]: DEBUG nova.compute.manager [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 965.112802] env[63372]: DEBUG nova.network.neutron [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 965.211406] env[63372]: DEBUG nova.policy [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7ffe6776b664c00b9086046f9171ec6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46d2cb188a004dc385f2aa15458be929', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 965.228428] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024563, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.331210] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72c9aac-8536-41e8-9931-65ad166ed9b4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.339584] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99207004-0709-401b-a74a-a5a88bc7cbaa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.372192] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47a175e-d242-4286-ae8a-758c2183e5c0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.380237] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f06bab-d3d1-46f9-8196-f40c50d4d89d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.393937] env[63372]: DEBUG nova.compute.provider_tree [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.483018] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Releasing lock "refresh_cache-6beadbab-2cc1-4b69-95c9-e1a0ea11045f" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.483710] env[63372]: DEBUG nova.compute.manager [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Instance network_info: |[{"id": "eef17039-cc41-41f1-9c94-291f90803267", "address": "fa:16:3e:87:d7:99", "network": {"id": "cbdcb945-f0a5-49b1-9b18-fcd8e4ca9728", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1068757189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e517915d5e7a4e3e924cc95c55170a66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeef17039-cc", "ovs_interfaceid": "eef17039-cc41-41f1-9c94-291f90803267", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 965.484202] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:d7:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd3c6b64-aba2-4bdc-a693-3b4dff3ed861', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eef17039-cc41-41f1-9c94-291f90803267', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 965.492527] env[63372]: DEBUG oslo.service.loopingcall [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.493644] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.493918] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 965.494167] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-16fcca8f-25c0-42aa-bbdb-9bafe5d32ae6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.514575] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.514575] env[63372]: value = "task-1024564" [ 965.514575] env[63372]: _type = "Task" [ 965.514575] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.524802] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024564, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.562784] env[63372]: DEBUG nova.network.neutron [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Successfully created port: 18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 965.615060] env[63372]: DEBUG nova.compute.manager [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 965.728976] env[63372]: DEBUG oslo_vmware.api [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024563, 'name': PowerOnVM_Task, 'duration_secs': 0.813683} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.729287] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 965.729498] env[63372]: INFO nova.compute.manager [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Took 7.21 seconds to spawn the instance on the hypervisor. [ 965.729684] env[63372]: DEBUG nova.compute.manager [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 965.730509] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bfb8351-cc61-402b-b9ee-ef73aac076bf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.787774] env[63372]: DEBUG nova.network.neutron [-] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.898241] env[63372]: DEBUG nova.scheduler.client.report [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.030316] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024564, 'name': CreateVM_Task, 'duration_secs': 0.357009} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.030316] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 966.030316] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.030316] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.030316] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.030316] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3a5a36f-f281-47af-a59e-2caa2c034b0f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.034571] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 966.034571] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e2f227-cafc-490f-6cae-bed9622cf0b3" [ 966.034571] env[63372]: _type = "Task" [ 966.034571] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.044234] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e2f227-cafc-490f-6cae-bed9622cf0b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.252047] env[63372]: INFO nova.compute.manager [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Took 23.13 seconds to build instance. [ 966.290503] env[63372]: INFO nova.compute.manager [-] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Took 1.36 seconds to deallocate network for instance. [ 966.404032] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.803s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.406499] env[63372]: DEBUG oslo_concurrency.lockutils [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.464s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.406873] env[63372]: DEBUG nova.objects.instance [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Lazy-loading 'resources' on Instance uuid c7c444cb-0aaa-4486-840a-e48fa255b421 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.443276] env[63372]: INFO nova.scheduler.client.report [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleted allocations for instance ba1daa0a-dcf8-4586-9964-a962fcc94196 [ 966.487198] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "93a5d948-0629-4f53-a681-858d519acfa7" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.487486] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "93a5d948-0629-4f53-a681-858d519acfa7" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.487684] env[63372]: DEBUG nova.compute.manager [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Going to confirm migration 3 {{(pid=63372) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 966.547686] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e2f227-cafc-490f-6cae-bed9622cf0b3, 'name': SearchDatastore_Task, 'duration_secs': 0.029062} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.548607] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.548800] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 966.549086] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.549252] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.549772] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.549772] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b175313c-c0cc-4e00-ba7b-a6c92079fe85 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.561910] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 966.562123] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 966.563861] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dece45e8-36ad-4a58-a8f0-db7e387501bd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.569090] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 966.569090] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52373a4c-242b-f613-25d0-0c996d26c651" [ 966.569090] env[63372]: _type = "Task" [ 966.569090] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.576145] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52373a4c-242b-f613-25d0-0c996d26c651, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.625530] env[63372]: DEBUG nova.compute.manager [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 966.656545] env[63372]: DEBUG nova.virt.hardware [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.656545] env[63372]: DEBUG nova.virt.hardware [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.656545] env[63372]: DEBUG nova.virt.hardware [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.656545] env[63372]: DEBUG nova.virt.hardware [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.656545] env[63372]: DEBUG nova.virt.hardware [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.656545] env[63372]: DEBUG nova.virt.hardware [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.656545] env[63372]: DEBUG nova.virt.hardware [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.656545] env[63372]: DEBUG nova.virt.hardware [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.656545] env[63372]: DEBUG nova.virt.hardware [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.656545] env[63372]: DEBUG nova.virt.hardware [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.656545] env[63372]: DEBUG nova.virt.hardware [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.657987] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1887a61-15a3-42fa-8b71-4249ef5ae0e9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.667112] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-774a5d47-1075-4ad7-8735-a7a08d558a1b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.753745] env[63372]: DEBUG oslo_concurrency.lockutils [None req-56fad5ae-1e97-421f-a647-9e9586dc26d9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "b7732621-7e58-40a3-b723-5c66df6f74a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.637s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.797665] env[63372]: DEBUG oslo_concurrency.lockutils [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.927216] env[63372]: DEBUG nova.compute.manager [req-c14556d7-2960-4c24-989a-5b90c0c43560 req-06f7ebe5-8391-48b7-9444-5ea3f908f843 service nova] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Received event network-vif-deleted-46264f2d-4aad-453f-b549-c48687e3b6ad {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.956978] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b3544003-51f0-47c1-808f-98ebc1252add tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "ba1daa0a-dcf8-4586-9964-a962fcc94196" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.528s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.067484] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.067673] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.067860] env[63372]: DEBUG nova.network.neutron [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 967.068267] env[63372]: DEBUG nova.objects.instance [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lazy-loading 'info_cache' on Instance uuid 93a5d948-0629-4f53-a681-858d519acfa7 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.080356] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52373a4c-242b-f613-25d0-0c996d26c651, 'name': SearchDatastore_Task, 'duration_secs': 0.033411} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.081206] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b50c2b0-9580-4b60-940c-ac454f0e8e16 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.086238] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 967.086238] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52163bb7-fd4e-7eb4-562a-51e4cb65705b" [ 967.086238] env[63372]: _type = "Task" [ 967.086238] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.091263] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f773fe62-3cd7-479f-a1e4-e0adbf12eaf9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.098832] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52163bb7-fd4e-7eb4-562a-51e4cb65705b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.102400] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c0aa59f-58ca-4d6b-810a-77b927aa4a74 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.135516] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ae3252-7a53-4343-8b9b-4a555382791a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.144519] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bafe588-cdca-421c-a33f-89c383ec87f1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.159554] env[63372]: DEBUG nova.compute.provider_tree [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.327902] env[63372]: DEBUG nova.network.neutron [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Successfully updated port: 18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 967.597799] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52163bb7-fd4e-7eb4-562a-51e4cb65705b, 'name': SearchDatastore_Task, 'duration_secs': 0.024126} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.597799] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.597990] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 6beadbab-2cc1-4b69-95c9-e1a0ea11045f/6beadbab-2cc1-4b69-95c9-e1a0ea11045f.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 967.598802] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7414a7c4-487a-4961-8306-3e301e58b25f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.605533] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 967.605533] env[63372]: value = "task-1024565" [ 967.605533] env[63372]: _type = "Task" [ 967.605533] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.613754] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024565, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.662801] env[63372]: DEBUG nova.scheduler.client.report [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.831460] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.831610] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.831768] env[63372]: DEBUG nova.network.neutron [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 968.119119] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024565, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.171365] env[63372]: DEBUG oslo_concurrency.lockutils [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.762s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.171365] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 6.097s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.171365] env[63372]: DEBUG nova.objects.instance [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63372) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 968.196884] env[63372]: INFO nova.scheduler.client.report [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Deleted allocations for instance c7c444cb-0aaa-4486-840a-e48fa255b421 [ 968.331921] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "a3145c98-e046-4b55-b80a-57e63807bd7a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.332351] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "a3145c98-e046-4b55-b80a-57e63807bd7a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.397666] env[63372]: DEBUG nova.network.neutron [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 968.405495] env[63372]: DEBUG nova.network.neutron [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance_info_cache with network_info: [{"id": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "address": "fa:16:3e:3f:99:77", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c4c74ba-ad", "ovs_interfaceid": "8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.620138] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024565, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.698645} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.620446] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 6beadbab-2cc1-4b69-95c9-e1a0ea11045f/6beadbab-2cc1-4b69-95c9-e1a0ea11045f.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 968.620668] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 968.620923] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-746140d8-73e5-4882-b1a7-a509b280eeed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.627491] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 968.627491] env[63372]: value = "task-1024566" [ 968.627491] env[63372]: _type = "Task" [ 968.627491] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.635378] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024566, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.653000] env[63372]: DEBUG nova.network.neutron [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updating instance_info_cache with network_info: [{"id": "18c7850a-d97e-4564-92e5-38d1204369e4", "address": "fa:16:3e:14:28:4a", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c7850a-d9", "ovs_interfaceid": "18c7850a-d97e-4564-92e5-38d1204369e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.705354] env[63372]: DEBUG oslo_concurrency.lockutils [None req-de42bc31-c292-4d1a-9a18-4d68e46badc8 tempest-ServerGroupTestJSON-1254723579 tempest-ServerGroupTestJSON-1254723579-project-member] Lock "c7c444cb-0aaa-4486-840a-e48fa255b421" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.660s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.836747] env[63372]: DEBUG nova.compute.manager [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 968.908370] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-93a5d948-0629-4f53-a681-858d519acfa7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.908711] env[63372]: DEBUG nova.objects.instance [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lazy-loading 'migration_context' on Instance uuid 93a5d948-0629-4f53-a681-858d519acfa7 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.954982] env[63372]: DEBUG nova.compute.manager [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Received event network-changed-128a0686-7ca8-427f-9cf7-c63a2f370e1f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.955187] env[63372]: DEBUG nova.compute.manager [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Refreshing instance network info cache due to event network-changed-128a0686-7ca8-427f-9cf7-c63a2f370e1f. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 968.955388] env[63372]: DEBUG oslo_concurrency.lockutils [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] Acquiring lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.955508] env[63372]: DEBUG oslo_concurrency.lockutils [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] Acquired lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.955672] env[63372]: DEBUG nova.network.neutron [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Refreshing network info cache for port 128a0686-7ca8-427f-9cf7-c63a2f370e1f {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 969.138425] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024566, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093555} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.138729] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 969.139525] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f7c11e-b874-49d5-a8f5-b13ea29522c6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.162136] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 6beadbab-2cc1-4b69-95c9-e1a0ea11045f/6beadbab-2cc1-4b69-95c9-e1a0ea11045f.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 969.162620] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.162896] env[63372]: DEBUG nova.compute.manager [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Instance network_info: |[{"id": "18c7850a-d97e-4564-92e5-38d1204369e4", "address": "fa:16:3e:14:28:4a", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c7850a-d9", "ovs_interfaceid": "18c7850a-d97e-4564-92e5-38d1204369e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 969.163153] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5dac511-770f-4524-bc09-3f949ec37feb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.178177] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:28:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '18c7850a-d97e-4564-92e5-38d1204369e4', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.185399] env[63372]: DEBUG oslo.service.loopingcall [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.186367] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ca01abba-f2a1-48fa-9e5c-fdda97e7917e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.187351] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 969.188278] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.412s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.188503] env[63372]: DEBUG nova.objects.instance [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lazy-loading 'resources' on Instance uuid 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.189384] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4a23567d-df3f-44ee-a883-601ac267c7dd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.206293] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 969.206293] env[63372]: value = "task-1024567" [ 969.206293] env[63372]: _type = "Task" [ 969.206293] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.211814] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.211814] env[63372]: value = "task-1024568" [ 969.211814] env[63372]: _type = "Task" [ 969.211814] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.221051] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024567, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.226351] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024568, 'name': CreateVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.358404] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.411694] env[63372]: DEBUG nova.objects.base [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Object Instance<93a5d948-0629-4f53-a681-858d519acfa7> lazy-loaded attributes: info_cache,migration_context {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 969.412278] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-965a34b1-7c9e-441a-a7e7-415526a41e32 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.435210] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f87b6d6-c2f1-4f67-ab48-d52f6598f1b4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.440428] env[63372]: DEBUG oslo_vmware.api [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 969.440428] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e60e12-d466-a0c5-62e4-6bc7d5c6af1d" [ 969.440428] env[63372]: _type = "Task" [ 969.440428] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.448310] env[63372]: DEBUG oslo_vmware.api [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e60e12-d466-a0c5-62e4-6bc7d5c6af1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.716547] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024567, 'name': ReconfigVM_Task, 'duration_secs': 0.370941} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.720450] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 6beadbab-2cc1-4b69-95c9-e1a0ea11045f/6beadbab-2cc1-4b69-95c9-e1a0ea11045f.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.721102] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f767ed20-fbd2-4e1b-a487-4bc319b7c90c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.729486] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024568, 'name': CreateVM_Task, 'duration_secs': 0.34983} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.730586] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 969.730919] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 969.730919] env[63372]: value = "task-1024569" [ 969.730919] env[63372]: _type = "Task" [ 969.730919] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.731644] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.731792] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.732114] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 969.732426] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8260a4e3-a71b-40dd-851a-9efcb85aa004 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.744485] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024569, 'name': Rename_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.745135] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 969.745135] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52a4df2e-bf8d-1088-1226-d21684ab49db" [ 969.745135] env[63372]: _type = "Task" [ 969.745135] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.753562] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a4df2e-bf8d-1088-1226-d21684ab49db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.883856] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8467eae-aa26-4294-b13f-54e71c955b13 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.892388] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa9d7bd-685d-4af5-9ccf-c3f6ce94f59c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.931085] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa1c0c04-032b-49ae-bec1-3eeb0875fbc0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.938915] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282e1ee5-59a7-46d5-b323-6a0d97f06db4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.955700] env[63372]: DEBUG nova.compute.provider_tree [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.960989] env[63372]: DEBUG oslo_vmware.api [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e60e12-d466-a0c5-62e4-6bc7d5c6af1d, 'name': SearchDatastore_Task, 'duration_secs': 0.012758} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.960989] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.963827] env[63372]: DEBUG nova.network.neutron [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updated VIF entry in instance network info cache for port 128a0686-7ca8-427f-9cf7-c63a2f370e1f. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 969.964169] env[63372]: DEBUG nova.network.neutron [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance_info_cache with network_info: [{"id": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "address": "fa:16:3e:28:02:b8", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap128a0686-7c", "ovs_interfaceid": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.243888] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024569, 'name': Rename_Task, 'duration_secs': 0.205044} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.244176] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 970.244420] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ce031bb-1c91-4661-98a7-598a6176e97c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.255118] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a4df2e-bf8d-1088-1226-d21684ab49db, 'name': SearchDatastore_Task, 'duration_secs': 0.023497} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.256250] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.256483] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 970.256781] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.256933] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.257132] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.257425] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 970.257425] env[63372]: value = "task-1024570" [ 970.257425] env[63372]: _type = "Task" [ 970.257425] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.257605] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b46fc2f-a622-4dd4-aa09-356eb5c0be0b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.267265] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.269711] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.269964] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 970.270909] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96cd0b8c-e1fb-4a13-ba6f-a7af0fa906d5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.276395] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 970.276395] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52a285f1-abdc-11ca-2e2a-44a2847476d4" [ 970.276395] env[63372]: _type = "Task" [ 970.276395] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.284669] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a285f1-abdc-11ca-2e2a-44a2847476d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.462334] env[63372]: DEBUG nova.scheduler.client.report [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.467358] env[63372]: DEBUG oslo_concurrency.lockutils [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] Releasing lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.467528] env[63372]: DEBUG nova.compute.manager [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Received event network-vif-plugged-18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.467804] env[63372]: DEBUG oslo_concurrency.lockutils [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] Acquiring lock "b455b16b-3332-43bf-a91e-a350287ba5f3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.468587] env[63372]: DEBUG oslo_concurrency.lockutils [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] Lock "b455b16b-3332-43bf-a91e-a350287ba5f3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.468587] env[63372]: DEBUG oslo_concurrency.lockutils [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] Lock "b455b16b-3332-43bf-a91e-a350287ba5f3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.468587] env[63372]: DEBUG nova.compute.manager [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] No waiting events found dispatching network-vif-plugged-18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 970.469358] env[63372]: WARNING nova.compute.manager [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Received unexpected event network-vif-plugged-18c7850a-d97e-4564-92e5-38d1204369e4 for instance with vm_state building and task_state spawning. [ 970.469358] env[63372]: DEBUG nova.compute.manager [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Received event network-changed-18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.469358] env[63372]: DEBUG nova.compute.manager [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Refreshing instance network info cache due to event network-changed-18c7850a-d97e-4564-92e5-38d1204369e4. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 970.469584] env[63372]: DEBUG oslo_concurrency.lockutils [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] Acquiring lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.469798] env[63372]: DEBUG oslo_concurrency.lockutils [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] Acquired lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.470047] env[63372]: DEBUG nova.network.neutron [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Refreshing network info cache for port 18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 970.769214] env[63372]: DEBUG oslo_vmware.api [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024570, 'name': PowerOnVM_Task, 'duration_secs': 0.427376} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.769499] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 970.769706] env[63372]: INFO nova.compute.manager [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Took 9.98 seconds to spawn the instance on the hypervisor. [ 970.769886] env[63372]: DEBUG nova.compute.manager [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.770694] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3e129b4-4aee-43f0-9f7c-f96247b618ce {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.787087] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a285f1-abdc-11ca-2e2a-44a2847476d4, 'name': SearchDatastore_Task, 'duration_secs': 0.013937} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.787893] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73a8da65-e9c0-41d9-a269-cf451328985e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.792781] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 970.792781] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]528ab401-f84d-6465-1107-20b5a598c4fa" [ 970.792781] env[63372]: _type = "Task" [ 970.792781] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.801916] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]528ab401-f84d-6465-1107-20b5a598c4fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.973080] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.785s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.977212] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.484s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.979198] env[63372]: INFO nova.compute.claims [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 971.003732] env[63372]: INFO nova.scheduler.client.report [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleted allocations for instance 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912 [ 971.290810] env[63372]: INFO nova.compute.manager [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Took 27.32 seconds to build instance. [ 971.302581] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]528ab401-f84d-6465-1107-20b5a598c4fa, 'name': SearchDatastore_Task, 'duration_secs': 0.009934} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.303506] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.303758] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] b455b16b-3332-43bf-a91e-a350287ba5f3/b455b16b-3332-43bf-a91e-a350287ba5f3.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 971.305581] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e2bcb5fc-1616-4e65-838e-964501822317 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.313653] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 971.313653] env[63372]: value = "task-1024571" [ 971.313653] env[63372]: _type = "Task" [ 971.313653] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.322126] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024571, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.514812] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cb4fd74a-149c-47a6-9690-a417f44d7f6b tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "4ee6c0ac-b7ac-49e7-b029-ca6e9a595912" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.282s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.556430] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "6beadbab-2cc1-4b69-95c9-e1a0ea11045f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.567698] env[63372]: DEBUG nova.network.neutron [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updated VIF entry in instance network info cache for port 18c7850a-d97e-4564-92e5-38d1204369e4. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 971.568359] env[63372]: DEBUG nova.network.neutron [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updating instance_info_cache with network_info: [{"id": "18c7850a-d97e-4564-92e5-38d1204369e4", "address": "fa:16:3e:14:28:4a", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c7850a-d9", "ovs_interfaceid": "18c7850a-d97e-4564-92e5-38d1204369e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.791148] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1a690d9e-64d3-4ad3-a8c4-a75124e0d6af tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "6beadbab-2cc1-4b69-95c9-e1a0ea11045f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.829s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.791610] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "6beadbab-2cc1-4b69-95c9-e1a0ea11045f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.235s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.791845] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "6beadbab-2cc1-4b69-95c9-e1a0ea11045f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.792083] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "6beadbab-2cc1-4b69-95c9-e1a0ea11045f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.792321] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "6beadbab-2cc1-4b69-95c9-e1a0ea11045f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.794791] env[63372]: INFO nova.compute.manager [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Terminating instance [ 971.796792] env[63372]: DEBUG nova.compute.manager [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 971.796991] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 971.797875] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77816e75-fb4a-43c1-abb9-5fe8eaa99998 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.805408] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 971.805645] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1bf4d41-9c29-4181-829a-306b125be8c2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.811612] env[63372]: DEBUG oslo_vmware.api [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 971.811612] env[63372]: value = "task-1024572" [ 971.811612] env[63372]: _type = "Task" [ 971.811612] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.822020] env[63372]: DEBUG oslo_vmware.api [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024572, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.826790] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024571, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.071756] env[63372]: DEBUG oslo_concurrency.lockutils [req-322085c3-4fdd-4f63-9d78-a2f8518966d4 req-3fe2053b-c3fa-4d64-a9d8-89cf5b7b8723 service nova] Releasing lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.164385] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-097eafc5-28de-40e7-a7a5-ff8526608e38 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.171805] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43676c96-647e-42b1-a716-0c7c822e9476 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.204132] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c967c0-ae13-455c-bddc-2188c3e74753 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.211861] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83de84f0-f417-4d8d-8888-947f64ade7e0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.226483] env[63372]: DEBUG nova.compute.provider_tree [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.323531] env[63372]: DEBUG oslo_vmware.api [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024572, 'name': PowerOffVM_Task, 'duration_secs': 0.168256} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.326620] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 972.326794] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 972.327056] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024571, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.530804} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.327260] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c38e842d-40d8-40e0-b62e-662039f63add {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.328607] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] b455b16b-3332-43bf-a91e-a350287ba5f3/b455b16b-3332-43bf-a91e-a350287ba5f3.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 972.328867] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 972.329044] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-076e806e-33fe-4179-856f-3883065bd639 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.336633] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 972.336633] env[63372]: value = "task-1024574" [ 972.336633] env[63372]: _type = "Task" [ 972.336633] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.344682] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024574, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.393900] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 972.394153] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 972.394337] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Deleting the datastore file [datastore1] 6beadbab-2cc1-4b69-95c9-e1a0ea11045f {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.394711] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-757799b3-ff7e-4459-bb48-752fda292742 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.400792] env[63372]: DEBUG oslo_vmware.api [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 972.400792] env[63372]: value = "task-1024575" [ 972.400792] env[63372]: _type = "Task" [ 972.400792] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.409785] env[63372]: DEBUG oslo_vmware.api [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024575, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.731649] env[63372]: DEBUG nova.scheduler.client.report [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.788682] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "c01a5d24-eb46-4a69-993e-753880ce8e85" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.788968] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.848522] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024574, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.443341} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.848778] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 972.849642] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31065b9b-f0f0-4d7e-8c50-39fe9a11ce9c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.872546] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] b455b16b-3332-43bf-a91e-a350287ba5f3/b455b16b-3332-43bf-a91e-a350287ba5f3.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 972.872546] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9f12529-3afd-4016-9531-731876e41756 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.892639] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 972.892639] env[63372]: value = "task-1024576" [ 972.892639] env[63372]: _type = "Task" [ 972.892639] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.900911] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024576, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.908646] env[63372]: DEBUG oslo_vmware.api [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024575, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.236900] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.260s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.237471] env[63372]: DEBUG nova.compute.manager [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 973.240684] env[63372]: DEBUG oslo_concurrency.lockutils [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.443s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.240917] env[63372]: DEBUG nova.objects.instance [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lazy-loading 'resources' on Instance uuid 61f7a052-9a42-4df3-a568-eab197fc5c4f {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.291812] env[63372]: INFO nova.compute.manager [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Detaching volume cccbd23a-5f23-47d3-9e07-0c6321e2cfe3 [ 973.339490] env[63372]: INFO nova.virt.block_device [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Attempting to driver detach volume cccbd23a-5f23-47d3-9e07-0c6321e2cfe3 from mountpoint /dev/sdb [ 973.339900] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Volume detach. Driver type: vmdk {{(pid=63372) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 973.340355] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227396', 'volume_id': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'name': 'volume-cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c01a5d24-eb46-4a69-993e-753880ce8e85', 'attached_at': '', 'detached_at': '', 'volume_id': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'serial': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 973.341789] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.342178] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.344810] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae35e0e-6409-4a7a-b263-5bfc1f76acd0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.370943] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20675f9-2835-404a-955c-5aeb24c0a9f0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.380044] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfde9d95-7fce-4be6-b015-b5a95253615d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.410615] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf03ca5-239b-4630-9710-24d6276a777f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.418324] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024576, 'name': ReconfigVM_Task, 'duration_secs': 0.289344} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.432763] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Reconfigured VM instance instance-0000005c to attach disk [datastore2] b455b16b-3332-43bf-a91e-a350287ba5f3/b455b16b-3332-43bf-a91e-a350287ba5f3.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.433730] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] The volume has not been displaced from its original location: [datastore1] volume-cccbd23a-5f23-47d3-9e07-0c6321e2cfe3/volume-cccbd23a-5f23-47d3-9e07-0c6321e2cfe3.vmdk. No consolidation needed. {{(pid=63372) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 973.439990] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Reconfiguring VM instance instance-00000041 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 973.440519] env[63372]: DEBUG oslo_vmware.api [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024575, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.580688} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.440881] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dbac8d05-c03a-49df-ae83-2c73da47a609 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.443092] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3a26fc8-a41f-4a8d-8380-9c6fd47de8a0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.456074] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.456451] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 973.456767] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 973.457166] env[63372]: INFO nova.compute.manager [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Took 1.66 seconds to destroy the instance on the hypervisor. [ 973.457669] env[63372]: DEBUG oslo.service.loopingcall [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.458609] env[63372]: DEBUG nova.compute.manager [-] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 973.458829] env[63372]: DEBUG nova.network.neutron [-] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 973.466208] env[63372]: DEBUG oslo_vmware.api [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 973.466208] env[63372]: value = "task-1024578" [ 973.466208] env[63372]: _type = "Task" [ 973.466208] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.467531] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 973.467531] env[63372]: value = "task-1024577" [ 973.467531] env[63372]: _type = "Task" [ 973.467531] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.478892] env[63372]: DEBUG oslo_vmware.api [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024578, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.482100] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024577, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.746052] env[63372]: DEBUG nova.compute.utils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 973.749663] env[63372]: DEBUG nova.compute.manager [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 973.749663] env[63372]: DEBUG nova.network.neutron [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 973.847024] env[63372]: DEBUG nova.compute.manager [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 973.894205] env[63372]: DEBUG nova.policy [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2253a9a409d4f1ba9176485655c0988', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dac773c2536745e285181426ae34bb96', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 973.984549] env[63372]: DEBUG oslo_vmware.api [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024578, 'name': ReconfigVM_Task, 'duration_secs': 0.366071} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.985150] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024577, 'name': Rename_Task, 'duration_secs': 0.165795} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.988269] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Reconfigured VM instance instance-00000041 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 973.993744] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 973.994950] env[63372]: DEBUG nova.compute.manager [req-00353c03-3b0b-4a97-b175-70e7c1629cdc req-7fd73b2d-38f6-4383-bd5b-68ba6aeda090 service nova] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Received event network-vif-deleted-eef17039-cc41-41f1-9c94-291f90803267 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 973.995161] env[63372]: INFO nova.compute.manager [req-00353c03-3b0b-4a97-b175-70e7c1629cdc req-7fd73b2d-38f6-4383-bd5b-68ba6aeda090 service nova] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Neutron deleted interface eef17039-cc41-41f1-9c94-291f90803267; detaching it from the instance and deleting it from the info cache [ 973.995337] env[63372]: DEBUG nova.network.neutron [req-00353c03-3b0b-4a97-b175-70e7c1629cdc req-7fd73b2d-38f6-4383-bd5b-68ba6aeda090 service nova] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.997364] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-121840d4-bb68-4d53-835d-9179f1a48173 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.007461] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b69a8a29-810a-488c-8533-ed5035fa0669 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.011515] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21533a12-3103-469d-86cb-84eb54cc2f02 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.016453] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 974.016453] env[63372]: value = "task-1024579" [ 974.016453] env[63372]: _type = "Task" [ 974.016453] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.018098] env[63372]: DEBUG oslo_vmware.api [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 974.018098] env[63372]: value = "task-1024580" [ 974.018098] env[63372]: _type = "Task" [ 974.018098] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.028068] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d94d2a80-28fe-4692-92dd-57ade3b57c03 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.037457] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024579, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.044215] env[63372]: DEBUG oslo_vmware.api [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024580, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.072286] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3eea80c-af11-412a-b091-316d64e0cca6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.079692] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e9e14f-306e-4aeb-8018-0a937dc62d9d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.093031] env[63372]: DEBUG nova.compute.provider_tree [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.260086] env[63372]: DEBUG nova.compute.manager [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 974.261435] env[63372]: DEBUG nova.network.neutron [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Successfully created port: 298d217f-c17a-4ee2-acae-baee355112dd {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 974.369693] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.436684] env[63372]: DEBUG nova.network.neutron [-] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.509701] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-beb31517-38e9-4b93-b764-e020fb75fb66 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.520305] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5332de95-7a17-4d38-ab97-353c556a8aba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.552723] env[63372]: DEBUG oslo_vmware.api [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024579, 'name': PowerOnVM_Task, 'duration_secs': 0.468718} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.556282] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 974.556756] env[63372]: INFO nova.compute.manager [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Took 7.93 seconds to spawn the instance on the hypervisor. [ 974.556756] env[63372]: DEBUG nova.compute.manager [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 974.557066] env[63372]: DEBUG oslo_vmware.api [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024580, 'name': ReconfigVM_Task, 'duration_secs': 0.20605} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.564608] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dcfe325-dc20-42e0-b416-0dfac92e998a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.567397] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227396', 'volume_id': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'name': 'volume-cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c01a5d24-eb46-4a69-993e-753880ce8e85', 'attached_at': '', 'detached_at': '', 'volume_id': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3', 'serial': 'cccbd23a-5f23-47d3-9e07-0c6321e2cfe3'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 974.569588] env[63372]: DEBUG nova.compute.manager [req-00353c03-3b0b-4a97-b175-70e7c1629cdc req-7fd73b2d-38f6-4383-bd5b-68ba6aeda090 service nova] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Detach interface failed, port_id=eef17039-cc41-41f1-9c94-291f90803267, reason: Instance 6beadbab-2cc1-4b69-95c9-e1a0ea11045f could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 974.596822] env[63372]: DEBUG nova.scheduler.client.report [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.939328] env[63372]: INFO nova.compute.manager [-] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Took 1.48 seconds to deallocate network for instance. [ 975.089929] env[63372]: INFO nova.compute.manager [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Took 22.61 seconds to build instance. [ 975.102236] env[63372]: DEBUG oslo_concurrency.lockutils [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.861s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.104777] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.746s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.107281] env[63372]: INFO nova.compute.claims [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 975.126315] env[63372]: DEBUG nova.objects.instance [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lazy-loading 'flavor' on Instance uuid c01a5d24-eb46-4a69-993e-753880ce8e85 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.131837] env[63372]: INFO nova.scheduler.client.report [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleted allocations for instance 61f7a052-9a42-4df3-a568-eab197fc5c4f [ 975.270867] env[63372]: DEBUG nova.compute.manager [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 975.293918] env[63372]: DEBUG nova.virt.hardware [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.294189] env[63372]: DEBUG nova.virt.hardware [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.294342] env[63372]: DEBUG nova.virt.hardware [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.294526] env[63372]: DEBUG nova.virt.hardware [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.294675] env[63372]: DEBUG nova.virt.hardware [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.294826] env[63372]: DEBUG nova.virt.hardware [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.295066] env[63372]: DEBUG nova.virt.hardware [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.295703] env[63372]: DEBUG nova.virt.hardware [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.295703] env[63372]: DEBUG nova.virt.hardware [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.295703] env[63372]: DEBUG nova.virt.hardware [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.295703] env[63372]: DEBUG nova.virt.hardware [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.296944] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748d5f1d-4990-481b-97d9-26f216ed614d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.304722] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3706f8e7-5edc-4c54-a3fe-131729527b33 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.363440] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.363796] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.447118] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.591878] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fe1c42ab-9686-425f-a9c7-3a5a56334526 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "b455b16b-3332-43bf-a91e-a350287ba5f3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.118s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.641639] env[63372]: DEBUG oslo_concurrency.lockutils [None req-12c28eac-0d36-4538-bce1-73e6987b001a tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "61f7a052-9a42-4df3-a568-eab197fc5c4f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.849s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.798642] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "1724ce03-c8e2-415d-a380-59ac69fbfb57" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.799613] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "1724ce03-c8e2-415d-a380-59ac69fbfb57" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.868528] env[63372]: DEBUG nova.compute.manager [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 976.135648] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0146e1ef-291f-4e20-bd4d-69673f2a0d73 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.346s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.306501] env[63372]: DEBUG nova.compute.manager [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 976.316778] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d371aae-d7fb-4991-9596-3d44fc9f545b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.326865] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "c01a5d24-eb46-4a69-993e-753880ce8e85" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.328196] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.328196] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "c01a5d24-eb46-4a69-993e-753880ce8e85-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.328196] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.328196] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.331464] env[63372]: INFO nova.compute.manager [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Terminating instance [ 976.336231] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2562b160-7111-40e5-9b8a-bf42e77164f7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.340922] env[63372]: DEBUG nova.compute.manager [req-2b8d2577-18db-489a-8820-30ed71c37487 req-37ebc736-3e56-46c1-9343-20138b4d6513 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Received event network-changed-18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.341153] env[63372]: DEBUG nova.compute.manager [req-2b8d2577-18db-489a-8820-30ed71c37487 req-37ebc736-3e56-46c1-9343-20138b4d6513 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Refreshing instance network info cache due to event network-changed-18c7850a-d97e-4564-92e5-38d1204369e4. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 976.341912] env[63372]: DEBUG oslo_concurrency.lockutils [req-2b8d2577-18db-489a-8820-30ed71c37487 req-37ebc736-3e56-46c1-9343-20138b4d6513 service nova] Acquiring lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.341912] env[63372]: DEBUG oslo_concurrency.lockutils [req-2b8d2577-18db-489a-8820-30ed71c37487 req-37ebc736-3e56-46c1-9343-20138b4d6513 service nova] Acquired lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.341912] env[63372]: DEBUG nova.network.neutron [req-2b8d2577-18db-489a-8820-30ed71c37487 req-37ebc736-3e56-46c1-9343-20138b4d6513 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Refreshing network info cache for port 18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 976.344632] env[63372]: DEBUG nova.compute.manager [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 976.344632] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 976.345296] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d910b7a4-4fc7-45a3-8905-7aac8b516254 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.394495] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69bb0a9b-f221-4225-861c-8d63a1e9aa94 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.394495] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 976.398180] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a53564d-deb3-41a2-b7d0-1c10a3cedf3c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.405319] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ef4ead-0011-4c7f-a132-5ab73d308a1c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.411575] env[63372]: DEBUG oslo_vmware.api [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 976.411575] env[63372]: value = "task-1024581" [ 976.411575] env[63372]: _type = "Task" [ 976.411575] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.427021] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.427021] env[63372]: DEBUG nova.compute.provider_tree [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.434136] env[63372]: DEBUG oslo_vmware.api [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024581, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.561197] env[63372]: DEBUG nova.network.neutron [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Successfully updated port: 298d217f-c17a-4ee2-acae-baee355112dd {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 976.824596] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.919299] env[63372]: DEBUG oslo_vmware.api [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024581, 'name': PowerOffVM_Task, 'duration_secs': 0.239558} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.919562] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 976.919729] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 976.919966] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc61b717-fcdd-4ee1-b1d8-73400172fef6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.929424] env[63372]: DEBUG nova.scheduler.client.report [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.982939] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 976.983245] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 976.983489] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleting the datastore file [datastore2] c01a5d24-eb46-4a69-993e-753880ce8e85 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 976.986452] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b4b54351-b998-4b6c-965a-249beef5ffc8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.993365] env[63372]: DEBUG oslo_vmware.api [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 976.993365] env[63372]: value = "task-1024583" [ 976.993365] env[63372]: _type = "Task" [ 976.993365] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.001765] env[63372]: DEBUG oslo_vmware.api [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024583, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.064040] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.064040] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.064040] env[63372]: DEBUG nova.network.neutron [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 977.114658] env[63372]: DEBUG nova.network.neutron [req-2b8d2577-18db-489a-8820-30ed71c37487 req-37ebc736-3e56-46c1-9343-20138b4d6513 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updated VIF entry in instance network info cache for port 18c7850a-d97e-4564-92e5-38d1204369e4. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 977.115050] env[63372]: DEBUG nova.network.neutron [req-2b8d2577-18db-489a-8820-30ed71c37487 req-37ebc736-3e56-46c1-9343-20138b4d6513 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updating instance_info_cache with network_info: [{"id": "18c7850a-d97e-4564-92e5-38d1204369e4", "address": "fa:16:3e:14:28:4a", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c7850a-d9", "ovs_interfaceid": "18c7850a-d97e-4564-92e5-38d1204369e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.298894] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.299372] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.434031] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.329s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.434524] env[63372]: DEBUG nova.compute.manager [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 977.438096] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 7.477s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.506743] env[63372]: DEBUG oslo_vmware.api [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024583, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13602} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.507015] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 977.507217] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 977.507392] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 977.507567] env[63372]: INFO nova.compute.manager [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Took 1.16 seconds to destroy the instance on the hypervisor. [ 977.507806] env[63372]: DEBUG oslo.service.loopingcall [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.507995] env[63372]: DEBUG nova.compute.manager [-] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 977.508121] env[63372]: DEBUG nova.network.neutron [-] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 977.620825] env[63372]: DEBUG oslo_concurrency.lockutils [req-2b8d2577-18db-489a-8820-30ed71c37487 req-37ebc736-3e56-46c1-9343-20138b4d6513 service nova] Releasing lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.620825] env[63372]: DEBUG nova.network.neutron [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 977.801355] env[63372]: DEBUG nova.compute.manager [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 977.871201] env[63372]: DEBUG nova.network.neutron [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating instance_info_cache with network_info: [{"id": "298d217f-c17a-4ee2-acae-baee355112dd", "address": "fa:16:3e:66:af:86", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298d217f-c1", "ovs_interfaceid": "298d217f-c17a-4ee2-acae-baee355112dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.941680] env[63372]: DEBUG nova.compute.utils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 977.946027] env[63372]: DEBUG nova.compute.manager [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 977.946027] env[63372]: DEBUG nova.network.neutron [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 978.016323] env[63372]: DEBUG nova.policy [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb233d4b6fe54fefae3046d62d167e34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d0c0f315749429dbae22a72ae82b1b2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 978.119510] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a222617-c26e-4aa0-8370-3d403cb02b38 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.126878] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d89ef7-7eef-4858-938d-caace08aceaa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.158740] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a56f25fb-bb34-45e4-9378-b110290d323b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.166119] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c86fa496-30f5-4405-b389-7051a02ed182 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.180800] env[63372]: DEBUG nova.compute.provider_tree [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.322270] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.374634] env[63372]: DEBUG nova.compute.manager [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Received event network-vif-plugged-298d217f-c17a-4ee2-acae-baee355112dd {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.374937] env[63372]: DEBUG oslo_concurrency.lockutils [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] Acquiring lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.375914] env[63372]: DEBUG oslo_concurrency.lockutils [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] Lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.375914] env[63372]: DEBUG oslo_concurrency.lockutils [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] Lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.375914] env[63372]: DEBUG nova.compute.manager [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] No waiting events found dispatching network-vif-plugged-298d217f-c17a-4ee2-acae-baee355112dd {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 978.376295] env[63372]: WARNING nova.compute.manager [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Received unexpected event network-vif-plugged-298d217f-c17a-4ee2-acae-baee355112dd for instance with vm_state building and task_state spawning. [ 978.376295] env[63372]: DEBUG nova.compute.manager [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Received event network-changed-298d217f-c17a-4ee2-acae-baee355112dd {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.376636] env[63372]: DEBUG nova.compute.manager [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Refreshing instance network info cache due to event network-changed-298d217f-c17a-4ee2-acae-baee355112dd. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 978.376636] env[63372]: DEBUG oslo_concurrency.lockutils [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] Acquiring lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.376920] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.378377] env[63372]: DEBUG nova.compute.manager [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Instance network_info: |[{"id": "298d217f-c17a-4ee2-acae-baee355112dd", "address": "fa:16:3e:66:af:86", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298d217f-c1", "ovs_interfaceid": "298d217f-c17a-4ee2-acae-baee355112dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 978.378377] env[63372]: DEBUG oslo_concurrency.lockutils [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] Acquired lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.378377] env[63372]: DEBUG nova.network.neutron [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Refreshing network info cache for port 298d217f-c17a-4ee2-acae-baee355112dd {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 978.379560] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:af:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d4ef133-b6f3-41d1-add4-92a1482195cf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '298d217f-c17a-4ee2-acae-baee355112dd', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 978.389410] env[63372]: DEBUG oslo.service.loopingcall [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.392655] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 978.393482] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6d9c5982-0439-4109-a0f3-0da3a83eb4a2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.417418] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 978.417418] env[63372]: value = "task-1024584" [ 978.417418] env[63372]: _type = "Task" [ 978.417418] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.426724] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024584, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.448545] env[63372]: DEBUG nova.compute.manager [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 978.521760] env[63372]: DEBUG nova.network.neutron [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Successfully created port: 6f03ed68-f858-4584-a059-ec0f2fa8e9d6 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 978.686024] env[63372]: DEBUG nova.scheduler.client.report [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.766988] env[63372]: DEBUG nova.network.neutron [-] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.929589] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024584, 'name': CreateVM_Task, 'duration_secs': 0.419344} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.929757] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 978.930505] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.930669] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.930981] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 978.931445] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afefd32d-d588-4d8d-89db-77df3792764b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.935432] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 978.935432] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5284c52e-e71d-ac83-2923-19261c1d6fb8" [ 978.935432] env[63372]: _type = "Task" [ 978.935432] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.946023] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5284c52e-e71d-ac83-2923-19261c1d6fb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.081614] env[63372]: DEBUG nova.network.neutron [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updated VIF entry in instance network info cache for port 298d217f-c17a-4ee2-acae-baee355112dd. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 979.082049] env[63372]: DEBUG nova.network.neutron [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating instance_info_cache with network_info: [{"id": "298d217f-c17a-4ee2-acae-baee355112dd", "address": "fa:16:3e:66:af:86", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298d217f-c1", "ovs_interfaceid": "298d217f-c17a-4ee2-acae-baee355112dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.269804] env[63372]: INFO nova.compute.manager [-] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Took 1.76 seconds to deallocate network for instance. [ 979.445719] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5284c52e-e71d-ac83-2923-19261c1d6fb8, 'name': SearchDatastore_Task, 'duration_secs': 0.010808} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.446030] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.446272] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.446508] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.446670] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.446852] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.447118] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a2e598b-97ef-41b3-9413-2842c3cf3ca1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.455690] env[63372]: DEBUG nova.compute.manager [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 979.457542] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.457726] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 979.458677] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e0c8222-c3fa-4dbe-afdd-a952499756d0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.464980] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 979.464980] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520839a7-1ab0-c618-f018-c6e00b9b4001" [ 979.464980] env[63372]: _type = "Task" [ 979.464980] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.472500] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520839a7-1ab0-c618-f018-c6e00b9b4001, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.483606] env[63372]: DEBUG nova.virt.hardware [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.483853] env[63372]: DEBUG nova.virt.hardware [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.484173] env[63372]: DEBUG nova.virt.hardware [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.484255] env[63372]: DEBUG nova.virt.hardware [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.484368] env[63372]: DEBUG nova.virt.hardware [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.484521] env[63372]: DEBUG nova.virt.hardware [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.484733] env[63372]: DEBUG nova.virt.hardware [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.484907] env[63372]: DEBUG nova.virt.hardware [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.485097] env[63372]: DEBUG nova.virt.hardware [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.485281] env[63372]: DEBUG nova.virt.hardware [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.485462] env[63372]: DEBUG nova.virt.hardware [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.486271] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381ee348-46b5-42ad-a795-7cf26aa325a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.493199] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaacb5c8-f4de-4ee4-9df1-a32d923cb5ae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.584736] env[63372]: DEBUG oslo_concurrency.lockutils [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] Releasing lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.585047] env[63372]: DEBUG nova.compute.manager [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Received event network-vif-deleted-c92e8cc9-dc72-4f20-b087-1d323c502108 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 979.585243] env[63372]: INFO nova.compute.manager [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Neutron deleted interface c92e8cc9-dc72-4f20-b087-1d323c502108; detaching it from the instance and deleting it from the info cache [ 979.585427] env[63372]: DEBUG nova.network.neutron [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.695249] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.257s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.698235] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.329s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.699724] env[63372]: INFO nova.compute.claims [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 979.776316] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.976327] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520839a7-1ab0-c618-f018-c6e00b9b4001, 'name': SearchDatastore_Task, 'duration_secs': 0.008347} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.977372] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0b83b1b-8b39-4357-b3dc-7737e95d6293 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.982312] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 979.982312] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521a4424-3b4a-200f-e54e-ae852a65e54d" [ 979.982312] env[63372]: _type = "Task" [ 979.982312] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.990138] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521a4424-3b4a-200f-e54e-ae852a65e54d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.088510] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-50226712-7e55-46c0-a9be-b3852750910e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.097611] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeba0047-3e53-4d67-9226-db5797e70a57 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.123423] env[63372]: DEBUG nova.compute.manager [req-bcd8c00f-740c-4a00-9f96-ba991c8b345d req-7b09ddd2-98fa-44f5-bc42-191bdb58ba5e service nova] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Detach interface failed, port_id=c92e8cc9-dc72-4f20-b087-1d323c502108, reason: Instance c01a5d24-eb46-4a69-993e-753880ce8e85 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 980.167967] env[63372]: DEBUG nova.network.neutron [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Successfully updated port: 6f03ed68-f858-4584-a059-ec0f2fa8e9d6 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 980.260617] env[63372]: INFO nova.scheduler.client.report [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleted allocation for migration 4130afc1-bde5-42b0-a496-4513ec7d4f56 [ 980.400768] env[63372]: DEBUG nova.compute.manager [req-d8fac369-9868-41b0-9bf1-4ece4005d15a req-94dbcb1a-a7ff-4518-9102-5096018e52f1 service nova] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Received event network-vif-plugged-6f03ed68-f858-4584-a059-ec0f2fa8e9d6 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.400998] env[63372]: DEBUG oslo_concurrency.lockutils [req-d8fac369-9868-41b0-9bf1-4ece4005d15a req-94dbcb1a-a7ff-4518-9102-5096018e52f1 service nova] Acquiring lock "a3145c98-e046-4b55-b80a-57e63807bd7a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.401294] env[63372]: DEBUG oslo_concurrency.lockutils [req-d8fac369-9868-41b0-9bf1-4ece4005d15a req-94dbcb1a-a7ff-4518-9102-5096018e52f1 service nova] Lock "a3145c98-e046-4b55-b80a-57e63807bd7a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.401505] env[63372]: DEBUG oslo_concurrency.lockutils [req-d8fac369-9868-41b0-9bf1-4ece4005d15a req-94dbcb1a-a7ff-4518-9102-5096018e52f1 service nova] Lock "a3145c98-e046-4b55-b80a-57e63807bd7a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.401681] env[63372]: DEBUG nova.compute.manager [req-d8fac369-9868-41b0-9bf1-4ece4005d15a req-94dbcb1a-a7ff-4518-9102-5096018e52f1 service nova] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] No waiting events found dispatching network-vif-plugged-6f03ed68-f858-4584-a059-ec0f2fa8e9d6 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 980.401850] env[63372]: WARNING nova.compute.manager [req-d8fac369-9868-41b0-9bf1-4ece4005d15a req-94dbcb1a-a7ff-4518-9102-5096018e52f1 service nova] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Received unexpected event network-vif-plugged-6f03ed68-f858-4584-a059-ec0f2fa8e9d6 for instance with vm_state building and task_state spawning. [ 980.402029] env[63372]: DEBUG nova.compute.manager [req-d8fac369-9868-41b0-9bf1-4ece4005d15a req-94dbcb1a-a7ff-4518-9102-5096018e52f1 service nova] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Received event network-changed-6f03ed68-f858-4584-a059-ec0f2fa8e9d6 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.402198] env[63372]: DEBUG nova.compute.manager [req-d8fac369-9868-41b0-9bf1-4ece4005d15a req-94dbcb1a-a7ff-4518-9102-5096018e52f1 service nova] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Refreshing instance network info cache due to event network-changed-6f03ed68-f858-4584-a059-ec0f2fa8e9d6. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 980.402384] env[63372]: DEBUG oslo_concurrency.lockutils [req-d8fac369-9868-41b0-9bf1-4ece4005d15a req-94dbcb1a-a7ff-4518-9102-5096018e52f1 service nova] Acquiring lock "refresh_cache-a3145c98-e046-4b55-b80a-57e63807bd7a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.402547] env[63372]: DEBUG oslo_concurrency.lockutils [req-d8fac369-9868-41b0-9bf1-4ece4005d15a req-94dbcb1a-a7ff-4518-9102-5096018e52f1 service nova] Acquired lock "refresh_cache-a3145c98-e046-4b55-b80a-57e63807bd7a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.402717] env[63372]: DEBUG nova.network.neutron [req-d8fac369-9868-41b0-9bf1-4ece4005d15a req-94dbcb1a-a7ff-4518-9102-5096018e52f1 service nova] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Refreshing network info cache for port 6f03ed68-f858-4584-a059-ec0f2fa8e9d6 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 980.495042] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521a4424-3b4a-200f-e54e-ae852a65e54d, 'name': SearchDatastore_Task, 'duration_secs': 0.010619} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.495340] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.495605] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] cbb471c5-8d86-4d8f-8935-f6a77e5144be/cbb471c5-8d86-4d8f-8935-f6a77e5144be.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 980.495863] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-002f55ad-3236-4c1b-90dd-194fadf9fa5f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.502203] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 980.502203] env[63372]: value = "task-1024585" [ 980.502203] env[63372]: _type = "Task" [ 980.502203] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.511552] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024585, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.670695] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "refresh_cache-a3145c98-e046-4b55-b80a-57e63807bd7a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.767151] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cd9746aa-6900-4cf2-862a-56ebebd5481a tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "93a5d948-0629-4f53-a681-858d519acfa7" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 14.279s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.913779] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7705c124-f2a2-42e0-86e7-bee8e6f871dc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.923883] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a429f87-c34d-4d75-8f61-0cfb4997b4d6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.961611] env[63372]: DEBUG nova.network.neutron [req-d8fac369-9868-41b0-9bf1-4ece4005d15a req-94dbcb1a-a7ff-4518-9102-5096018e52f1 service nova] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 980.964124] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa1476dc-1dd3-4ea1-8a24-2f47419fc98a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.972248] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d3a96e-6b38-4054-8eaa-cbcae63f25d5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.985997] env[63372]: DEBUG nova.compute.provider_tree [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.012441] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024585, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461915} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.012748] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] cbb471c5-8d86-4d8f-8935-f6a77e5144be/cbb471c5-8d86-4d8f-8935-f6a77e5144be.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 981.012964] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 981.013232] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-568a256a-b034-4cfa-98a9-b435ede9e8ea {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.019847] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 981.019847] env[63372]: value = "task-1024586" [ 981.019847] env[63372]: _type = "Task" [ 981.019847] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.029334] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024586, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.090683] env[63372]: DEBUG nova.network.neutron [req-d8fac369-9868-41b0-9bf1-4ece4005d15a req-94dbcb1a-a7ff-4518-9102-5096018e52f1 service nova] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.288938] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "93a5d948-0629-4f53-a681-858d519acfa7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.289232] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "93a5d948-0629-4f53-a681-858d519acfa7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.289449] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "93a5d948-0629-4f53-a681-858d519acfa7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.289637] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "93a5d948-0629-4f53-a681-858d519acfa7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.289807] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "93a5d948-0629-4f53-a681-858d519acfa7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.292298] env[63372]: INFO nova.compute.manager [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Terminating instance [ 981.294124] env[63372]: DEBUG nova.compute.manager [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 981.294350] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 981.295186] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25fbbd9-9b9b-4337-961b-1e6117b5f94d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.303134] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 981.303378] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-04788ff1-08ac-4e4c-8bb8-34e4a95c9034 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.308686] env[63372]: DEBUG oslo_vmware.api [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 981.308686] env[63372]: value = "task-1024587" [ 981.308686] env[63372]: _type = "Task" [ 981.308686] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.316316] env[63372]: DEBUG oslo_vmware.api [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024587, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.489319] env[63372]: DEBUG nova.scheduler.client.report [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 981.529192] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024586, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06252} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.529566] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 981.530370] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91123e82-eb7b-4b68-83f4-18bc7dc9721a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.552414] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] cbb471c5-8d86-4d8f-8935-f6a77e5144be/cbb471c5-8d86-4d8f-8935-f6a77e5144be.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 981.552702] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41a11e4a-c97b-4373-947a-fd385509c993 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.572306] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 981.572306] env[63372]: value = "task-1024588" [ 981.572306] env[63372]: _type = "Task" [ 981.572306] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.580037] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024588, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.593699] env[63372]: DEBUG oslo_concurrency.lockutils [req-d8fac369-9868-41b0-9bf1-4ece4005d15a req-94dbcb1a-a7ff-4518-9102-5096018e52f1 service nova] Releasing lock "refresh_cache-a3145c98-e046-4b55-b80a-57e63807bd7a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.594124] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "refresh_cache-a3145c98-e046-4b55-b80a-57e63807bd7a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.594321] env[63372]: DEBUG nova.network.neutron [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 981.819420] env[63372]: DEBUG oslo_vmware.api [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024587, 'name': PowerOffVM_Task, 'duration_secs': 0.22125} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.819699] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 981.819870] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 981.820134] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a751e31f-b049-4c69-99b2-45aa5da591e9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.894235] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 981.894480] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 981.894672] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleting the datastore file [datastore2] 93a5d948-0629-4f53-a681-858d519acfa7 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.894938] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb1a61a3-4743-4f91-ab32-b946cbaab081 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.901804] env[63372]: DEBUG oslo_vmware.api [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 981.901804] env[63372]: value = "task-1024590" [ 981.901804] env[63372]: _type = "Task" [ 981.901804] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.909745] env[63372]: DEBUG oslo_vmware.api [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024590, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.994855] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.995415] env[63372]: DEBUG nova.compute.manager [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 981.997986] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.552s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.998284] env[63372]: DEBUG nova.objects.instance [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lazy-loading 'resources' on Instance uuid 6beadbab-2cc1-4b69-95c9-e1a0ea11045f {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.082331] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024588, 'name': ReconfigVM_Task, 'duration_secs': 0.256735} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.082621] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Reconfigured VM instance instance-0000005d to attach disk [datastore1] cbb471c5-8d86-4d8f-8935-f6a77e5144be/cbb471c5-8d86-4d8f-8935-f6a77e5144be.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 982.083241] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cced7fda-9f2b-4f11-9f0e-b3a6e76c5dd6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.088847] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 982.088847] env[63372]: value = "task-1024591" [ 982.088847] env[63372]: _type = "Task" [ 982.088847] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.097670] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024591, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.126505] env[63372]: DEBUG nova.network.neutron [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 982.303580] env[63372]: DEBUG nova.network.neutron [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Updating instance_info_cache with network_info: [{"id": "6f03ed68-f858-4584-a059-ec0f2fa8e9d6", "address": "fa:16:3e:e3:5c:77", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f03ed68-f8", "ovs_interfaceid": "6f03ed68-f858-4584-a059-ec0f2fa8e9d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.414058] env[63372]: DEBUG oslo_vmware.api [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024590, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142025} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.414392] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 982.414600] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 982.414813] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 982.415282] env[63372]: INFO nova.compute.manager [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Took 1.12 seconds to destroy the instance on the hypervisor. [ 982.415389] env[63372]: DEBUG oslo.service.loopingcall [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.415571] env[63372]: DEBUG nova.compute.manager [-] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 982.415669] env[63372]: DEBUG nova.network.neutron [-] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 982.501648] env[63372]: DEBUG nova.compute.utils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 982.506327] env[63372]: DEBUG nova.compute.manager [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 982.506494] env[63372]: DEBUG nova.network.neutron [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 982.572504] env[63372]: DEBUG nova.policy [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '118d2f1de2be45cc8bb48bb75525d37e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'afe39cfedf214d50be775ef736f94da9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 982.599093] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024591, 'name': Rename_Task, 'duration_secs': 0.142017} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.599360] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 982.599627] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ba222ed4-340b-4db3-82ab-f63d52c21912 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.605793] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 982.605793] env[63372]: value = "task-1024592" [ 982.605793] env[63372]: _type = "Task" [ 982.605793] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.616423] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024592, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.665838] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebeb5ce-98b7-4228-b72d-dbe9caf57377 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.675209] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80711a95-ed99-47a1-a61d-46bc3c79434a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.706236] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ecb160-07f0-4ad7-98d7-fdcde9544af0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.713742] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7fe9fbe-b40a-47c8-a22e-663285690f10 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.732356] env[63372]: DEBUG nova.compute.provider_tree [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 982.806116] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "refresh_cache-a3145c98-e046-4b55-b80a-57e63807bd7a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.806571] env[63372]: DEBUG nova.compute.manager [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Instance network_info: |[{"id": "6f03ed68-f858-4584-a059-ec0f2fa8e9d6", "address": "fa:16:3e:e3:5c:77", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f03ed68-f8", "ovs_interfaceid": "6f03ed68-f858-4584-a059-ec0f2fa8e9d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 982.807511] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:5c:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60badc2d-69d2-467d-a92e-98511f5cb0b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6f03ed68-f858-4584-a059-ec0f2fa8e9d6', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.815352] env[63372]: DEBUG oslo.service.loopingcall [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.815595] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 982.815819] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d002d7ee-aac3-4ded-9b2a-230572a3c286 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.835780] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.835780] env[63372]: value = "task-1024593" [ 982.835780] env[63372]: _type = "Task" [ 982.835780] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.845529] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024593, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.006948] env[63372]: DEBUG nova.compute.manager [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 983.113580] env[63372]: DEBUG nova.network.neutron [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Successfully created port: 9c53b112-6225-4e7e-95d5-cce3c462a615 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 983.124553] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024592, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.224826] env[63372]: DEBUG nova.compute.manager [req-980029bf-e70f-46be-a326-591010a2ac47 req-ef985d9e-2db7-49a0-ab58-4619a1912732 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Received event network-vif-deleted-8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.224826] env[63372]: INFO nova.compute.manager [req-980029bf-e70f-46be-a326-591010a2ac47 req-ef985d9e-2db7-49a0-ab58-4619a1912732 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Neutron deleted interface 8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f; detaching it from the instance and deleting it from the info cache [ 983.225125] env[63372]: DEBUG nova.network.neutron [req-980029bf-e70f-46be-a326-591010a2ac47 req-ef985d9e-2db7-49a0-ab58-4619a1912732 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.258570] env[63372]: ERROR nova.scheduler.client.report [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [req-ea837476-3b0b-46cc-a43e-16281f74f5f1] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ea837476-3b0b-46cc-a43e-16281f74f5f1"}]} [ 983.275613] env[63372]: DEBUG nova.scheduler.client.report [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Refreshing inventories for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 983.289733] env[63372]: DEBUG nova.scheduler.client.report [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Updating ProviderTree inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 983.290012] env[63372]: DEBUG nova.compute.provider_tree [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 983.304060] env[63372]: DEBUG nova.scheduler.client.report [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Refreshing aggregate associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, aggregates: None {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 983.324934] env[63372]: DEBUG nova.scheduler.client.report [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Refreshing trait associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 983.347806] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024593, 'name': CreateVM_Task, 'duration_secs': 0.385933} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.347995] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 983.352898] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.353110] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.353456] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 983.353736] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92bc1169-3e60-4301-a2be-e495710a1980 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.359333] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 983.359333] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52d03326-d5c8-fbde-7248-39634cdb660d" [ 983.359333] env[63372]: _type = "Task" [ 983.359333] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.371375] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52d03326-d5c8-fbde-7248-39634cdb660d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.514825] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae939c9-55a5-43e2-8b24-0a315d3cebe8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.522416] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604c937a-254f-440b-83f8-3de35bcc4b94 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.555483] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd143d5-7569-4034-a157-9ae606cb5afe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.563614] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b71ea574-e502-48a8-922a-5de72fdbc934 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.579314] env[63372]: DEBUG nova.compute.provider_tree [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 983.617806] env[63372]: DEBUG oslo_vmware.api [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024592, 'name': PowerOnVM_Task, 'duration_secs': 0.825924} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.618103] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 983.618374] env[63372]: INFO nova.compute.manager [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Took 8.35 seconds to spawn the instance on the hypervisor. [ 983.618565] env[63372]: DEBUG nova.compute.manager [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.619532] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df39033-95d6-4ff0-a994-8266203df52a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.672190] env[63372]: DEBUG nova.network.neutron [-] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.727595] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cca34639-6720-48cb-bf6e-0777f79e75e6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.737395] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf7a93f-8998-47e3-a9c4-3c7b8149fb2b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.764474] env[63372]: DEBUG nova.compute.manager [req-980029bf-e70f-46be-a326-591010a2ac47 req-ef985d9e-2db7-49a0-ab58-4619a1912732 service nova] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Detach interface failed, port_id=8c4c74ba-ad30-4b7d-a7af-e1e9e78ceb6f, reason: Instance 93a5d948-0629-4f53-a681-858d519acfa7 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 983.869979] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52d03326-d5c8-fbde-7248-39634cdb660d, 'name': SearchDatastore_Task, 'duration_secs': 0.009957} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.872826] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.872826] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.872826] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.872826] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.872826] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.872826] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dfcf158f-feda-4654-9d3f-fcc1b0dd4b0e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.884124] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.885043] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 983.885043] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aceec735-ee45-4b90-a92c-3ca609d78989 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.890344] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 983.890344] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52968711-fc17-573d-02a1-ee1a696b1061" [ 983.890344] env[63372]: _type = "Task" [ 983.890344] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.897775] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52968711-fc17-573d-02a1-ee1a696b1061, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.019227] env[63372]: DEBUG nova.compute.manager [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 984.044542] env[63372]: DEBUG nova.virt.hardware [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 984.044798] env[63372]: DEBUG nova.virt.hardware [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 984.044961] env[63372]: DEBUG nova.virt.hardware [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.045166] env[63372]: DEBUG nova.virt.hardware [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 984.045320] env[63372]: DEBUG nova.virt.hardware [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.045471] env[63372]: DEBUG nova.virt.hardware [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 984.045675] env[63372]: DEBUG nova.virt.hardware [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 984.045834] env[63372]: DEBUG nova.virt.hardware [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 984.046036] env[63372]: DEBUG nova.virt.hardware [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 984.046173] env[63372]: DEBUG nova.virt.hardware [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 984.046347] env[63372]: DEBUG nova.virt.hardware [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.047249] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c18a5d-686a-4686-aa29-56b5aea71ecb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.055197] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6f2521-0574-4ed0-bfad-5cfa0f7d42d8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.109007] env[63372]: DEBUG nova.scheduler.client.report [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 132 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 984.109348] env[63372]: DEBUG nova.compute.provider_tree [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 132 to 133 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 984.109535] env[63372]: DEBUG nova.compute.provider_tree [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 984.136927] env[63372]: INFO nova.compute.manager [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Took 18.67 seconds to build instance. [ 984.177066] env[63372]: INFO nova.compute.manager [-] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Took 1.76 seconds to deallocate network for instance. [ 984.401019] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52968711-fc17-573d-02a1-ee1a696b1061, 'name': SearchDatastore_Task, 'duration_secs': 0.052762} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.401859] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df85e535-f648-458b-a2f2-09be75a5b4b2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.407531] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 984.407531] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525e7946-f9b6-874a-39ba-e659be83ab0a" [ 984.407531] env[63372]: _type = "Task" [ 984.407531] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.415208] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525e7946-f9b6-874a-39ba-e659be83ab0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.615620] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.617s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.618827] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.194s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.621378] env[63372]: INFO nova.compute.claims [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 984.639811] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f88f3e7-a925-4b47-a033-3bbe8dde9404 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.185s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.642970] env[63372]: INFO nova.scheduler.client.report [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Deleted allocations for instance 6beadbab-2cc1-4b69-95c9-e1a0ea11045f [ 984.683522] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.918014] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525e7946-f9b6-874a-39ba-e659be83ab0a, 'name': SearchDatastore_Task, 'duration_secs': 0.035629} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.918416] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.918671] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] a3145c98-e046-4b55-b80a-57e63807bd7a/a3145c98-e046-4b55-b80a-57e63807bd7a.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 984.918943] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f81d369-7d13-4744-abe0-3146f7dda4d6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.925649] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 984.925649] env[63372]: value = "task-1024594" [ 984.925649] env[63372]: _type = "Task" [ 984.925649] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.934653] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024594, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.155771] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9c1e15a8-3186-4aa7-8d86-1edfa21c3fc9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "6beadbab-2cc1-4b69-95c9-e1a0ea11045f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.363s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.272259] env[63372]: DEBUG nova.network.neutron [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Successfully updated port: 9c53b112-6225-4e7e-95d5-cce3c462a615 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 985.346669] env[63372]: DEBUG nova.compute.manager [req-c8109fe9-b98c-4810-86fa-6b759a554f6d req-cbb6a4c1-cf9b-4266-b7a9-f7c8791189c5 service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Received event network-vif-plugged-9c53b112-6225-4e7e-95d5-cce3c462a615 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.347034] env[63372]: DEBUG oslo_concurrency.lockutils [req-c8109fe9-b98c-4810-86fa-6b759a554f6d req-cbb6a4c1-cf9b-4266-b7a9-f7c8791189c5 service nova] Acquiring lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.348400] env[63372]: DEBUG oslo_concurrency.lockutils [req-c8109fe9-b98c-4810-86fa-6b759a554f6d req-cbb6a4c1-cf9b-4266-b7a9-f7c8791189c5 service nova] Lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.348400] env[63372]: DEBUG oslo_concurrency.lockutils [req-c8109fe9-b98c-4810-86fa-6b759a554f6d req-cbb6a4c1-cf9b-4266-b7a9-f7c8791189c5 service nova] Lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.348400] env[63372]: DEBUG nova.compute.manager [req-c8109fe9-b98c-4810-86fa-6b759a554f6d req-cbb6a4c1-cf9b-4266-b7a9-f7c8791189c5 service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] No waiting events found dispatching network-vif-plugged-9c53b112-6225-4e7e-95d5-cce3c462a615 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 985.348400] env[63372]: WARNING nova.compute.manager [req-c8109fe9-b98c-4810-86fa-6b759a554f6d req-cbb6a4c1-cf9b-4266-b7a9-f7c8791189c5 service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Received unexpected event network-vif-plugged-9c53b112-6225-4e7e-95d5-cce3c462a615 for instance with vm_state building and task_state spawning. [ 985.436798] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024594, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.776810] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "refresh_cache-9be70e66-2d81-4c7b-8dcc-e69815faef1a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.777081] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "refresh_cache-9be70e66-2d81-4c7b-8dcc-e69815faef1a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.777337] env[63372]: DEBUG nova.network.neutron [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 985.815714] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cbc968d-e892-4d99-91e1-6db23814dcc0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.826421] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a57cd141-eea5-49fc-a499-b7216b70fb6b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.860402] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c5ba47-5805-43b0-837b-4156ab4f2653 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.869375] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-585538e8-6789-4090-acd2-b68500b8c507 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.883389] env[63372]: DEBUG nova.compute.provider_tree [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.937802] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024594, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511278} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.938760] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] a3145c98-e046-4b55-b80a-57e63807bd7a/a3145c98-e046-4b55-b80a-57e63807bd7a.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 985.938760] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 985.938946] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-686d3e08-3563-4356-82c1-756651b211c6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.946781] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 985.946781] env[63372]: value = "task-1024595" [ 985.946781] env[63372]: _type = "Task" [ 985.946781] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.955806] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024595, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.269062] env[63372]: DEBUG nova.compute.manager [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Stashing vm_state: active {{(pid=63372) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 986.318623] env[63372]: DEBUG nova.network.neutron [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 986.387014] env[63372]: DEBUG nova.scheduler.client.report [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.455992] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024595, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.137917} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.456287] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 986.457084] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d353a20e-428e-487f-9d69-47b818c812f6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.480443] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] a3145c98-e046-4b55-b80a-57e63807bd7a/a3145c98-e046-4b55-b80a-57e63807bd7a.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.481473] env[63372]: DEBUG nova.network.neutron [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Updating instance_info_cache with network_info: [{"id": "9c53b112-6225-4e7e-95d5-cce3c462a615", "address": "fa:16:3e:70:bf:65", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c53b112-62", "ovs_interfaceid": "9c53b112-6225-4e7e-95d5-cce3c462a615", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.482619] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3320c1f1-fbb1-415a-83e5-df42439f2858 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.498715] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "refresh_cache-9be70e66-2d81-4c7b-8dcc-e69815faef1a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.499006] env[63372]: DEBUG nova.compute.manager [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Instance network_info: |[{"id": "9c53b112-6225-4e7e-95d5-cce3c462a615", "address": "fa:16:3e:70:bf:65", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c53b112-62", "ovs_interfaceid": "9c53b112-6225-4e7e-95d5-cce3c462a615", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 986.499410] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:bf:65', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b21ab10-d886-4453-9472-9e11fb3c450d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9c53b112-6225-4e7e-95d5-cce3c462a615', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 986.506658] env[63372]: DEBUG oslo.service.loopingcall [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.507563] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 986.507811] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3e351bc6-da30-43e6-9778-29da94542e7b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.523815] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 986.523815] env[63372]: value = "task-1024596" [ 986.523815] env[63372]: _type = "Task" [ 986.523815] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.528755] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 986.528755] env[63372]: value = "task-1024597" [ 986.528755] env[63372]: _type = "Task" [ 986.528755] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.532029] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024596, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.539660] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024597, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.792836] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.892898] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.274s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.893636] env[63372]: DEBUG nova.compute.manager [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 986.896224] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.072s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.897866] env[63372]: INFO nova.compute.claims [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 987.034641] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024596, 'name': ReconfigVM_Task, 'duration_secs': 0.443055} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.038052] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Reconfigured VM instance instance-0000005e to attach disk [datastore1] a3145c98-e046-4b55-b80a-57e63807bd7a/a3145c98-e046-4b55-b80a-57e63807bd7a.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.038460] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3e2cdeea-6fac-4f28-8203-085bd56122be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.044747] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024597, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.045946] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 987.045946] env[63372]: value = "task-1024598" [ 987.045946] env[63372]: _type = "Task" [ 987.045946] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.053251] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024598, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.404995] env[63372]: DEBUG nova.compute.utils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 987.408217] env[63372]: DEBUG nova.compute.manager [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 987.408477] env[63372]: DEBUG nova.network.neutron [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 987.420627] env[63372]: DEBUG nova.compute.manager [req-07ce89ef-c108-4eea-9b6b-650c10a02263 req-97f229aa-1545-4a68-a2ea-95d20205c17e service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Received event network-changed-9c53b112-6225-4e7e-95d5-cce3c462a615 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.420893] env[63372]: DEBUG nova.compute.manager [req-07ce89ef-c108-4eea-9b6b-650c10a02263 req-97f229aa-1545-4a68-a2ea-95d20205c17e service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Refreshing instance network info cache due to event network-changed-9c53b112-6225-4e7e-95d5-cce3c462a615. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 987.421178] env[63372]: DEBUG oslo_concurrency.lockutils [req-07ce89ef-c108-4eea-9b6b-650c10a02263 req-97f229aa-1545-4a68-a2ea-95d20205c17e service nova] Acquiring lock "refresh_cache-9be70e66-2d81-4c7b-8dcc-e69815faef1a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.421293] env[63372]: DEBUG oslo_concurrency.lockutils [req-07ce89ef-c108-4eea-9b6b-650c10a02263 req-97f229aa-1545-4a68-a2ea-95d20205c17e service nova] Acquired lock "refresh_cache-9be70e66-2d81-4c7b-8dcc-e69815faef1a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.421457] env[63372]: DEBUG nova.network.neutron [req-07ce89ef-c108-4eea-9b6b-650c10a02263 req-97f229aa-1545-4a68-a2ea-95d20205c17e service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Refreshing network info cache for port 9c53b112-6225-4e7e-95d5-cce3c462a615 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 987.479570] env[63372]: DEBUG nova.policy [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc9dec0b991f4f61a674a2a844b5462c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ca1f09f7f9e4b2b8010e478202373ea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 987.545279] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024597, 'name': CreateVM_Task, 'duration_secs': 0.869629} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.545454] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 987.546170] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.546350] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.546690] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 987.546997] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9656990-b2c3-4f6a-9b03-1a68930dc6c9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.554824] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 987.554824] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f2b37b-9f65-f433-159c-dad135dab818" [ 987.554824] env[63372]: _type = "Task" [ 987.554824] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.557861] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024598, 'name': Rename_Task, 'duration_secs': 0.135104} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.560798] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 987.561057] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ef50d54-5d17-4ec6-b5e7-d4d589a5664a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.568400] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f2b37b-9f65-f433-159c-dad135dab818, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.569590] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 987.569590] env[63372]: value = "task-1024599" [ 987.569590] env[63372]: _type = "Task" [ 987.569590] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.576491] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024599, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.743620] env[63372]: DEBUG nova.network.neutron [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Successfully created port: b54d2aaa-a37b-4db7-b735-f9d2a6a9534b {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 987.913197] env[63372]: DEBUG nova.compute.manager [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 988.070325] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f2b37b-9f65-f433-159c-dad135dab818, 'name': SearchDatastore_Task, 'duration_secs': 0.022909} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.073903] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.074215] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.074472] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.074626] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.074833] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.077573] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cceeeed9-3ef1-4db4-80cf-f04a24e4e1fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.086992] env[63372]: DEBUG oslo_vmware.api [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024599, 'name': PowerOnVM_Task, 'duration_secs': 0.507962} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.087939] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 988.088174] env[63372]: INFO nova.compute.manager [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Took 8.63 seconds to spawn the instance on the hypervisor. [ 988.088372] env[63372]: DEBUG nova.compute.manager [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.088657] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.088817] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 988.090714] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847450eb-a17d-4b07-b0bb-87b6dd5e5f7c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.093037] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6dfc97c5-1df3-44d6-93eb-fb01011fd0f9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.095652] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3288d3a8-9889-47b3-a576-e4e900734d74 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.100948] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 988.100948] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]526642ed-ec4b-4e66-a6bf-695ac7184073" [ 988.100948] env[63372]: _type = "Task" [ 988.100948] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.109688] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3865b734-2117-4764-8de4-690bcdf828a6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.117623] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526642ed-ec4b-4e66-a6bf-695ac7184073, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.145054] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8447ff37-27f3-493d-a8b4-04c6dbd3efc6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.152709] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbcac489-9ffe-4317-8ffd-27cf2a2728df {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.167938] env[63372]: DEBUG nova.compute.provider_tree [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.213081] env[63372]: DEBUG nova.network.neutron [req-07ce89ef-c108-4eea-9b6b-650c10a02263 req-97f229aa-1545-4a68-a2ea-95d20205c17e service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Updated VIF entry in instance network info cache for port 9c53b112-6225-4e7e-95d5-cce3c462a615. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 988.213495] env[63372]: DEBUG nova.network.neutron [req-07ce89ef-c108-4eea-9b6b-650c10a02263 req-97f229aa-1545-4a68-a2ea-95d20205c17e service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Updating instance_info_cache with network_info: [{"id": "9c53b112-6225-4e7e-95d5-cce3c462a615", "address": "fa:16:3e:70:bf:65", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c53b112-62", "ovs_interfaceid": "9c53b112-6225-4e7e-95d5-cce3c462a615", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.290233] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.290506] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.621028] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526642ed-ec4b-4e66-a6bf-695ac7184073, 'name': SearchDatastore_Task, 'duration_secs': 0.015995} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.624362] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3b7bdea-303f-46fb-88d4-a7ec5bd631f1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.627409] env[63372]: INFO nova.compute.manager [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Took 19.29 seconds to build instance. [ 988.632090] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 988.632090] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52d88796-285b-ea5e-feee-145196773f7f" [ 988.632090] env[63372]: _type = "Task" [ 988.632090] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.641788] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52d88796-285b-ea5e-feee-145196773f7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.670731] env[63372]: DEBUG nova.scheduler.client.report [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.716228] env[63372]: DEBUG oslo_concurrency.lockutils [req-07ce89ef-c108-4eea-9b6b-650c10a02263 req-97f229aa-1545-4a68-a2ea-95d20205c17e service nova] Releasing lock "refresh_cache-9be70e66-2d81-4c7b-8dcc-e69815faef1a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.800232] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.800407] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Starting heal instance info cache {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 988.925964] env[63372]: DEBUG nova.compute.manager [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 988.954170] env[63372]: DEBUG nova.virt.hardware [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 988.954445] env[63372]: DEBUG nova.virt.hardware [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 988.954610] env[63372]: DEBUG nova.virt.hardware [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.954796] env[63372]: DEBUG nova.virt.hardware [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 988.954947] env[63372]: DEBUG nova.virt.hardware [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.955121] env[63372]: DEBUG nova.virt.hardware [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 988.955332] env[63372]: DEBUG nova.virt.hardware [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 988.955492] env[63372]: DEBUG nova.virt.hardware [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 988.955660] env[63372]: DEBUG nova.virt.hardware [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 988.955913] env[63372]: DEBUG nova.virt.hardware [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 988.955998] env[63372]: DEBUG nova.virt.hardware [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 988.956937] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fc0f89-db9a-4f25-a683-fdb215eb5771 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.965748] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a046f6-4c98-484c-8501-170595c42ded {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.132706] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d9557484-d640-48e3-b0f3-79b18f123a26 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "a3145c98-e046-4b55-b80a-57e63807bd7a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.800s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.146586] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52d88796-285b-ea5e-feee-145196773f7f, 'name': SearchDatastore_Task, 'duration_secs': 0.011571} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.146924] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.147650] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 9be70e66-2d81-4c7b-8dcc-e69815faef1a/9be70e66-2d81-4c7b-8dcc-e69815faef1a.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 989.147650] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8370b07f-db51-4c4e-b5f0-b75e64a610a2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.154349] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 989.154349] env[63372]: value = "task-1024600" [ 989.154349] env[63372]: _type = "Task" [ 989.154349] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.162435] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024600, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.177797] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.281s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.178434] env[63372]: DEBUG nova.compute.manager [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 989.181048] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.859s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.186727] env[63372]: INFO nova.compute.claims [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 989.302021] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "a3145c98-e046-4b55-b80a-57e63807bd7a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.302314] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "a3145c98-e046-4b55-b80a-57e63807bd7a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.302529] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "a3145c98-e046-4b55-b80a-57e63807bd7a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.302733] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "a3145c98-e046-4b55-b80a-57e63807bd7a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.302916] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "a3145c98-e046-4b55-b80a-57e63807bd7a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.306217] env[63372]: INFO nova.compute.manager [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Terminating instance [ 989.308340] env[63372]: DEBUG nova.compute.manager [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 989.308537] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 989.309407] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bdcc99d-2227-48c3-87d5-3ef428ea7784 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.318942] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 989.318942] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47fe8600-812a-417c-8efd-7ced33210ed0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.325464] env[63372]: DEBUG oslo_vmware.api [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 989.325464] env[63372]: value = "task-1024601" [ 989.325464] env[63372]: _type = "Task" [ 989.325464] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.335368] env[63372]: DEBUG oslo_vmware.api [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024601, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.408742] env[63372]: DEBUG nova.network.neutron [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Successfully updated port: b54d2aaa-a37b-4db7-b735-f9d2a6a9534b {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 989.460691] env[63372]: DEBUG nova.compute.manager [req-dbecf128-2a4a-469b-90e1-a89aab42b1fd req-f53ccf69-15b8-4f43-bf2b-d4ca0527ec0f service nova] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Received event network-vif-plugged-b54d2aaa-a37b-4db7-b735-f9d2a6a9534b {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.460932] env[63372]: DEBUG oslo_concurrency.lockutils [req-dbecf128-2a4a-469b-90e1-a89aab42b1fd req-f53ccf69-15b8-4f43-bf2b-d4ca0527ec0f service nova] Acquiring lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.461196] env[63372]: DEBUG oslo_concurrency.lockutils [req-dbecf128-2a4a-469b-90e1-a89aab42b1fd req-f53ccf69-15b8-4f43-bf2b-d4ca0527ec0f service nova] Lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.461334] env[63372]: DEBUG oslo_concurrency.lockutils [req-dbecf128-2a4a-469b-90e1-a89aab42b1fd req-f53ccf69-15b8-4f43-bf2b-d4ca0527ec0f service nova] Lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.461617] env[63372]: DEBUG nova.compute.manager [req-dbecf128-2a4a-469b-90e1-a89aab42b1fd req-f53ccf69-15b8-4f43-bf2b-d4ca0527ec0f service nova] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] No waiting events found dispatching network-vif-plugged-b54d2aaa-a37b-4db7-b735-f9d2a6a9534b {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 989.461822] env[63372]: WARNING nova.compute.manager [req-dbecf128-2a4a-469b-90e1-a89aab42b1fd req-f53ccf69-15b8-4f43-bf2b-d4ca0527ec0f service nova] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Received unexpected event network-vif-plugged-b54d2aaa-a37b-4db7-b735-f9d2a6a9534b for instance with vm_state building and task_state spawning. [ 989.461989] env[63372]: DEBUG nova.compute.manager [req-dbecf128-2a4a-469b-90e1-a89aab42b1fd req-f53ccf69-15b8-4f43-bf2b-d4ca0527ec0f service nova] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Received event network-changed-b54d2aaa-a37b-4db7-b735-f9d2a6a9534b {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.462163] env[63372]: DEBUG nova.compute.manager [req-dbecf128-2a4a-469b-90e1-a89aab42b1fd req-f53ccf69-15b8-4f43-bf2b-d4ca0527ec0f service nova] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Refreshing instance network info cache due to event network-changed-b54d2aaa-a37b-4db7-b735-f9d2a6a9534b. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 989.462354] env[63372]: DEBUG oslo_concurrency.lockutils [req-dbecf128-2a4a-469b-90e1-a89aab42b1fd req-f53ccf69-15b8-4f43-bf2b-d4ca0527ec0f service nova] Acquiring lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.462492] env[63372]: DEBUG oslo_concurrency.lockutils [req-dbecf128-2a4a-469b-90e1-a89aab42b1fd req-f53ccf69-15b8-4f43-bf2b-d4ca0527ec0f service nova] Acquired lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.462649] env[63372]: DEBUG nova.network.neutron [req-dbecf128-2a4a-469b-90e1-a89aab42b1fd req-f53ccf69-15b8-4f43-bf2b-d4ca0527ec0f service nova] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Refreshing network info cache for port b54d2aaa-a37b-4db7-b735-f9d2a6a9534b {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 989.665058] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024600, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458984} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.665389] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 9be70e66-2d81-4c7b-8dcc-e69815faef1a/9be70e66-2d81-4c7b-8dcc-e69815faef1a.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 989.665648] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 989.665935] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4de6fbb6-d901-469f-aa56-725e4abeda2c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.672741] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 989.672741] env[63372]: value = "task-1024602" [ 989.672741] env[63372]: _type = "Task" [ 989.672741] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.681047] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024602, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.691602] env[63372]: DEBUG nova.compute.utils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 989.694398] env[63372]: DEBUG nova.compute.manager [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 989.694398] env[63372]: DEBUG nova.network.neutron [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 989.778410] env[63372]: DEBUG nova.policy [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc9dec0b991f4f61a674a2a844b5462c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ca1f09f7f9e4b2b8010e478202373ea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 989.837476] env[63372]: DEBUG oslo_vmware.api [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024601, 'name': PowerOffVM_Task, 'duration_secs': 0.174748} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.837746] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 989.837912] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 989.838181] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61c6f88a-e832-48d1-9a99-256c46a9188d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.859226] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "refresh_cache-c4718797-aa86-4ec0-94d3-6480bd6aa898" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.859378] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquired lock "refresh_cache-c4718797-aa86-4ec0-94d3-6480bd6aa898" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.859527] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Forcefully refreshing network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 989.912611] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.912918] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 989.913118] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 989.913299] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleting the datastore file [datastore1] a3145c98-e046-4b55-b80a-57e63807bd7a {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.913615] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2d0e8cf9-0f73-4209-9fd6-4750b2b662fe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.920825] env[63372]: DEBUG oslo_vmware.api [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 989.920825] env[63372]: value = "task-1024604" [ 989.920825] env[63372]: _type = "Task" [ 989.920825] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.928715] env[63372]: DEBUG oslo_vmware.api [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024604, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.035406] env[63372]: DEBUG nova.network.neutron [req-dbecf128-2a4a-469b-90e1-a89aab42b1fd req-f53ccf69-15b8-4f43-bf2b-d4ca0527ec0f service nova] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 990.182179] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024602, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061762} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.182483] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 990.183305] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e5aeeb-a492-488b-a2d6-4fd1e1b4036d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.198651] env[63372]: DEBUG nova.compute.manager [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 990.212802] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 9be70e66-2d81-4c7b-8dcc-e69815faef1a/9be70e66-2d81-4c7b-8dcc-e69815faef1a.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 990.215723] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43941073-0ef2-4e66-8c0d-75300b335326 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.238273] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 990.238273] env[63372]: value = "task-1024605" [ 990.238273] env[63372]: _type = "Task" [ 990.238273] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.244794] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024605, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.280039] env[63372]: DEBUG nova.network.neutron [req-dbecf128-2a4a-469b-90e1-a89aab42b1fd req-f53ccf69-15b8-4f43-bf2b-d4ca0527ec0f service nova] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.360537] env[63372]: DEBUG nova.network.neutron [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Successfully created port: 99872bc9-8786-4e6d-9978-42f880ba8407 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 990.435372] env[63372]: DEBUG oslo_vmware.api [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024604, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.441162] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba55f99c-22ea-4e45-ad52-3bc1c68c5a6f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.449801] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d13605-5b00-4fde-8817-f95112e999c4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.481057] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1236d028-01c4-4b6a-9b4d-092e202cf1df {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.488611] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8a3b71-994f-4701-9c48-c99b466c3c11 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.502427] env[63372]: DEBUG nova.compute.provider_tree [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.752830] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024605, 'name': ReconfigVM_Task, 'duration_secs': 0.286265} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.755638] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 9be70e66-2d81-4c7b-8dcc-e69815faef1a/9be70e66-2d81-4c7b-8dcc-e69815faef1a.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 990.756585] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-44ec2e63-d4a4-484e-a181-00aa5a53ce8c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.766637] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 990.766637] env[63372]: value = "task-1024606" [ 990.766637] env[63372]: _type = "Task" [ 990.766637] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.775560] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024606, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.784404] env[63372]: DEBUG oslo_concurrency.lockutils [req-dbecf128-2a4a-469b-90e1-a89aab42b1fd req-f53ccf69-15b8-4f43-bf2b-d4ca0527ec0f service nova] Releasing lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.784404] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.784404] env[63372]: DEBUG nova.network.neutron [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 990.935747] env[63372]: DEBUG oslo_vmware.api [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024604, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.66489} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.939558] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.939558] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 990.939558] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 990.939558] env[63372]: INFO nova.compute.manager [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Took 1.63 seconds to destroy the instance on the hypervisor. [ 990.939558] env[63372]: DEBUG oslo.service.loopingcall [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.939558] env[63372]: DEBUG nova.compute.manager [-] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 990.939558] env[63372]: DEBUG nova.network.neutron [-] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 991.009091] env[63372]: DEBUG nova.scheduler.client.report [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 991.222029] env[63372]: DEBUG nova.compute.manager [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 991.227410] env[63372]: DEBUG nova.compute.manager [req-d9a0836e-f075-4674-8131-bfdd31cc55a9 req-a223b6c3-8a4e-424c-ae21-09d95b9b4df3 service nova] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Received event network-vif-deleted-6f03ed68-f858-4584-a059-ec0f2fa8e9d6 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.227677] env[63372]: INFO nova.compute.manager [req-d9a0836e-f075-4674-8131-bfdd31cc55a9 req-a223b6c3-8a4e-424c-ae21-09d95b9b4df3 service nova] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Neutron deleted interface 6f03ed68-f858-4584-a059-ec0f2fa8e9d6; detaching it from the instance and deleting it from the info cache [ 991.227895] env[63372]: DEBUG nova.network.neutron [req-d9a0836e-f075-4674-8131-bfdd31cc55a9 req-a223b6c3-8a4e-424c-ae21-09d95b9b4df3 service nova] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.233280] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Updating instance_info_cache with network_info: [{"id": "fff45461-b9dc-4d60-989d-c7460bc0becb", "address": "fa:16:3e:77:91:d7", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfff45461-b9", "ovs_interfaceid": "fff45461-b9dc-4d60-989d-c7460bc0becb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.249708] env[63372]: DEBUG nova.virt.hardware [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 991.249971] env[63372]: DEBUG nova.virt.hardware [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 991.250148] env[63372]: DEBUG nova.virt.hardware [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.250332] env[63372]: DEBUG nova.virt.hardware [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 991.250537] env[63372]: DEBUG nova.virt.hardware [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.250701] env[63372]: DEBUG nova.virt.hardware [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 991.250910] env[63372]: DEBUG nova.virt.hardware [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 991.251110] env[63372]: DEBUG nova.virt.hardware [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 991.251472] env[63372]: DEBUG nova.virt.hardware [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 991.251472] env[63372]: DEBUG nova.virt.hardware [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 991.251635] env[63372]: DEBUG nova.virt.hardware [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 991.253011] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5529ec65-bdcb-44e2-937f-73a6b05e9e54 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.264077] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4c2026-2a2c-454e-83b1-93c566960597 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.285663] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024606, 'name': Rename_Task, 'duration_secs': 0.148292} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.287849] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 991.288150] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fc469602-d9e1-4a41-bf15-24a48be3e305 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.296409] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 991.296409] env[63372]: value = "task-1024607" [ 991.296409] env[63372]: _type = "Task" [ 991.296409] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.306234] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024607, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.320335] env[63372]: DEBUG nova.network.neutron [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 991.517748] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.336s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.518388] env[63372]: DEBUG nova.compute.manager [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 991.521484] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.745s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.525282] env[63372]: DEBUG nova.objects.instance [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lazy-loading 'resources' on Instance uuid c01a5d24-eb46-4a69-993e-753880ce8e85 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.537674] env[63372]: DEBUG nova.network.neutron [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Updating instance_info_cache with network_info: [{"id": "b54d2aaa-a37b-4db7-b735-f9d2a6a9534b", "address": "fa:16:3e:03:b6:cc", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb54d2aaa-a3", "ovs_interfaceid": "b54d2aaa-a37b-4db7-b735-f9d2a6a9534b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.703016] env[63372]: DEBUG nova.network.neutron [-] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.732747] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed2e6e35-7d33-4b5f-b801-aadf97872df4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.738220] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Releasing lock "refresh_cache-c4718797-aa86-4ec0-94d3-6480bd6aa898" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.738220] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Updated the network info_cache for instance {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 991.738329] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.738976] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.739303] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.740321] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.740321] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.740321] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.740321] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63372) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 991.740321] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager.update_available_resource {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.744990] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b36d0ac-c275-445d-ade3-098db64eae08 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.773169] env[63372]: DEBUG nova.compute.manager [req-d9a0836e-f075-4674-8131-bfdd31cc55a9 req-a223b6c3-8a4e-424c-ae21-09d95b9b4df3 service nova] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Detach interface failed, port_id=6f03ed68-f858-4584-a059-ec0f2fa8e9d6, reason: Instance a3145c98-e046-4b55-b80a-57e63807bd7a could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 991.806240] env[63372]: DEBUG oslo_vmware.api [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024607, 'name': PowerOnVM_Task, 'duration_secs': 0.46235} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.806532] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 991.806737] env[63372]: INFO nova.compute.manager [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Took 7.79 seconds to spawn the instance on the hypervisor. [ 991.806914] env[63372]: DEBUG nova.compute.manager [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 991.807734] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1efbbe68-624b-4080-835e-98b5e192b8f6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.028113] env[63372]: DEBUG nova.compute.utils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 992.032322] env[63372]: DEBUG nova.compute.manager [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 992.032601] env[63372]: DEBUG nova.network.neutron [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 992.040261] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.040562] env[63372]: DEBUG nova.compute.manager [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Instance network_info: |[{"id": "b54d2aaa-a37b-4db7-b735-f9d2a6a9534b", "address": "fa:16:3e:03:b6:cc", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb54d2aaa-a3", "ovs_interfaceid": "b54d2aaa-a37b-4db7-b735-f9d2a6a9534b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 992.041241] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:b6:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9f208df-1fb5-4403-9796-7fd19e4bfb85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b54d2aaa-a37b-4db7-b735-f9d2a6a9534b', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 992.049323] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Creating folder: Project (4ca1f09f7f9e4b2b8010e478202373ea). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 992.049947] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b5e22ab-0287-4c21-a148-4580e9688659 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.064167] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Created folder: Project (4ca1f09f7f9e4b2b8010e478202373ea) in parent group-v227230. [ 992.064362] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Creating folder: Instances. Parent ref: group-v227435. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 992.066809] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7408f50e-2124-4542-8675-7d7050995241 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.076219] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Created folder: Instances in parent group-v227435. [ 992.076486] env[63372]: DEBUG oslo.service.loopingcall [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.078779] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 992.079275] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-272eb714-8bea-4387-b064-64bd13ae3f68 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.101433] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 992.101433] env[63372]: value = "task-1024610" [ 992.101433] env[63372]: _type = "Task" [ 992.101433] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.112379] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024610, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.206277] env[63372]: INFO nova.compute.manager [-] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Took 1.27 seconds to deallocate network for instance. [ 992.207260] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d347c26-2d4f-4e99-a42c-87dc32dca185 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.218335] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5dee641-8cd9-421d-adb1-4caf20ad74a9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.256693] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcaf5c28-d88a-46c5-b852-4faebfda06f4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.263340] env[63372]: DEBUG nova.policy [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7ffe6776b664c00b9086046f9171ec6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46d2cb188a004dc385f2aa15458be929', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 992.264028] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.271759] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2407741-1c2b-43e9-acc4-c776a89db24a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.292470] env[63372]: DEBUG nova.compute.provider_tree [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.297324] env[63372]: DEBUG nova.network.neutron [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Successfully updated port: 99872bc9-8786-4e6d-9978-42f880ba8407 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 992.324799] env[63372]: INFO nova.compute.manager [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Took 17.98 seconds to build instance. [ 992.430270] env[63372]: DEBUG nova.compute.manager [req-f5843a35-f00b-4dda-874f-bda137d844a7 req-d988d65d-0459-4730-bede-d574c6185e4b service nova] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Received event network-vif-plugged-99872bc9-8786-4e6d-9978-42f880ba8407 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.430557] env[63372]: DEBUG oslo_concurrency.lockutils [req-f5843a35-f00b-4dda-874f-bda137d844a7 req-d988d65d-0459-4730-bede-d574c6185e4b service nova] Acquiring lock "1724ce03-c8e2-415d-a380-59ac69fbfb57-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.431207] env[63372]: DEBUG oslo_concurrency.lockutils [req-f5843a35-f00b-4dda-874f-bda137d844a7 req-d988d65d-0459-4730-bede-d574c6185e4b service nova] Lock "1724ce03-c8e2-415d-a380-59ac69fbfb57-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.431306] env[63372]: DEBUG oslo_concurrency.lockutils [req-f5843a35-f00b-4dda-874f-bda137d844a7 req-d988d65d-0459-4730-bede-d574c6185e4b service nova] Lock "1724ce03-c8e2-415d-a380-59ac69fbfb57-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.431564] env[63372]: DEBUG nova.compute.manager [req-f5843a35-f00b-4dda-874f-bda137d844a7 req-d988d65d-0459-4730-bede-d574c6185e4b service nova] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] No waiting events found dispatching network-vif-plugged-99872bc9-8786-4e6d-9978-42f880ba8407 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 992.431757] env[63372]: WARNING nova.compute.manager [req-f5843a35-f00b-4dda-874f-bda137d844a7 req-d988d65d-0459-4730-bede-d574c6185e4b service nova] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Received unexpected event network-vif-plugged-99872bc9-8786-4e6d-9978-42f880ba8407 for instance with vm_state building and task_state spawning. [ 992.535506] env[63372]: DEBUG nova.compute.manager [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 992.611591] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024610, 'name': CreateVM_Task, 'duration_secs': 0.346467} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.611800] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 992.612475] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.612654] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.612962] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 992.613249] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34910dbd-e889-4536-80e1-939426dc61c4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.617931] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 992.617931] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521b139f-d2ee-38b1-8833-3d4928723df6" [ 992.617931] env[63372]: _type = "Task" [ 992.617931] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.625927] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521b139f-d2ee-38b1-8833-3d4928723df6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.718812] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.768344] env[63372]: DEBUG nova.compute.manager [req-239eed41-ca79-45da-808e-0a2b56029512 req-5b0c5cac-afe4-4a87-89b7-08fc59331254 service nova] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Received event network-changed-99872bc9-8786-4e6d-9978-42f880ba8407 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.768846] env[63372]: DEBUG nova.compute.manager [req-239eed41-ca79-45da-808e-0a2b56029512 req-5b0c5cac-afe4-4a87-89b7-08fc59331254 service nova] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Refreshing instance network info cache due to event network-changed-99872bc9-8786-4e6d-9978-42f880ba8407. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 992.769105] env[63372]: DEBUG oslo_concurrency.lockutils [req-239eed41-ca79-45da-808e-0a2b56029512 req-5b0c5cac-afe4-4a87-89b7-08fc59331254 service nova] Acquiring lock "refresh_cache-1724ce03-c8e2-415d-a380-59ac69fbfb57" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.769255] env[63372]: DEBUG oslo_concurrency.lockutils [req-239eed41-ca79-45da-808e-0a2b56029512 req-5b0c5cac-afe4-4a87-89b7-08fc59331254 service nova] Acquired lock "refresh_cache-1724ce03-c8e2-415d-a380-59ac69fbfb57" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.769421] env[63372]: DEBUG nova.network.neutron [req-239eed41-ca79-45da-808e-0a2b56029512 req-5b0c5cac-afe4-4a87-89b7-08fc59331254 service nova] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Refreshing network info cache for port 99872bc9-8786-4e6d-9978-42f880ba8407 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 992.773872] env[63372]: DEBUG nova.network.neutron [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Successfully created port: 301c1926-329c-454a-9755-e33bef4550d8 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 992.793957] env[63372]: DEBUG nova.scheduler.client.report [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.801192] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "refresh_cache-1724ce03-c8e2-415d-a380-59ac69fbfb57" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.827280] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c6ee3f30-3a91-4ad7-b888-08d365c915a9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.485s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.132018] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521b139f-d2ee-38b1-8833-3d4928723df6, 'name': SearchDatastore_Task, 'duration_secs': 0.033819} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.132018] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.132018] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 993.132018] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.132018] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.132018] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 993.132018] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e4958538-5730-48cd-aeeb-cc0b3076f48e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.137899] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 993.138301] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 993.139081] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-002ed90f-fb76-4fbb-b17d-7df749bbecdb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.144077] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 993.144077] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52343484-4f0a-1881-97b8-c5723bcbc345" [ 993.144077] env[63372]: _type = "Task" [ 993.144077] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.152288] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52343484-4f0a-1881-97b8-c5723bcbc345, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.299846] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.778s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.303012] env[63372]: DEBUG nova.network.neutron [req-239eed41-ca79-45da-808e-0a2b56029512 req-5b0c5cac-afe4-4a87-89b7-08fc59331254 service nova] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 993.305027] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.621s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.305027] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.307023] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.514s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.324326] env[63372]: INFO nova.scheduler.client.report [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleted allocations for instance c01a5d24-eb46-4a69-993e-753880ce8e85 [ 993.331170] env[63372]: DEBUG oslo_concurrency.lockutils [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.331342] env[63372]: DEBUG oslo_concurrency.lockutils [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.331521] env[63372]: INFO nova.compute.manager [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Shelving [ 993.333418] env[63372]: INFO nova.scheduler.client.report [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleted allocations for instance 93a5d948-0629-4f53-a681-858d519acfa7 [ 993.461663] env[63372]: DEBUG nova.network.neutron [req-239eed41-ca79-45da-808e-0a2b56029512 req-5b0c5cac-afe4-4a87-89b7-08fc59331254 service nova] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.547629] env[63372]: DEBUG nova.compute.manager [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 993.577442] env[63372]: DEBUG nova.virt.hardware [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.577699] env[63372]: DEBUG nova.virt.hardware [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.577862] env[63372]: DEBUG nova.virt.hardware [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.578057] env[63372]: DEBUG nova.virt.hardware [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.578214] env[63372]: DEBUG nova.virt.hardware [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.578387] env[63372]: DEBUG nova.virt.hardware [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.578615] env[63372]: DEBUG nova.virt.hardware [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.578780] env[63372]: DEBUG nova.virt.hardware [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.578945] env[63372]: DEBUG nova.virt.hardware [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.579129] env[63372]: DEBUG nova.virt.hardware [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.579311] env[63372]: DEBUG nova.virt.hardware [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.580203] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748e4789-56c9-4269-8658-664e279221b5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.588692] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed8849a0-8510-4ced-be1b-46386f57e3fc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.654928] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52343484-4f0a-1881-97b8-c5723bcbc345, 'name': SearchDatastore_Task, 'duration_secs': 0.007578} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.655732] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f9d65b2-3e83-4ea4-92b2-b2c4cec5fbc8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.661051] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 993.661051] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523d812b-f374-8928-2ed8-bb802794f10b" [ 993.661051] env[63372]: _type = "Task" [ 993.661051] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.668362] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523d812b-f374-8928-2ed8-bb802794f10b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.812198] env[63372]: INFO nova.compute.claims [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.834792] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dde17fcb-6c8e-4372-a79e-9fba251d116e tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "c01a5d24-eb46-4a69-993e-753880ce8e85" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.508s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.844500] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e85fd43-da9b-43ae-86a0-7fc737e0a78f tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "93a5d948-0629-4f53-a681-858d519acfa7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.555s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.845706] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 993.845955] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fc8d98b-efb4-4e45-95db-d9db72c2eb1e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.853092] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 993.853092] env[63372]: value = "task-1024611" [ 993.853092] env[63372]: _type = "Task" [ 993.853092] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.863824] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024611, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.964759] env[63372]: DEBUG oslo_concurrency.lockutils [req-239eed41-ca79-45da-808e-0a2b56029512 req-5b0c5cac-afe4-4a87-89b7-08fc59331254 service nova] Releasing lock "refresh_cache-1724ce03-c8e2-415d-a380-59ac69fbfb57" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.965126] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "refresh_cache-1724ce03-c8e2-415d-a380-59ac69fbfb57" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.965295] env[63372]: DEBUG nova.network.neutron [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 994.173775] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523d812b-f374-8928-2ed8-bb802794f10b, 'name': SearchDatastore_Task, 'duration_secs': 0.009261} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.173775] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.174070] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] ddca7aa5-fcc8-4835-bfdb-47781335f2cd/ddca7aa5-fcc8-4835-bfdb-47781335f2cd.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 994.174285] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f482d219-e8a3-4adc-8240-8d73cf05085f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.180564] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 994.180564] env[63372]: value = "task-1024612" [ 994.180564] env[63372]: _type = "Task" [ 994.180564] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.188669] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024612, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.319511] env[63372]: INFO nova.compute.resource_tracker [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating resource usage from migration 75fa60a2-d2ae-4672-8d3c-8a09a4cdd49a [ 994.371158] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024611, 'name': PowerOffVM_Task, 'duration_secs': 0.191734} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.371745] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 994.373069] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df391eb7-becc-4433-915f-b5ec7007b382 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.399281] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d339e2c4-21ad-426a-a315-5dd65f36d5cd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.543477] env[63372]: DEBUG nova.network.neutron [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 994.551529] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742276c4-2ac1-441a-a444-7221d7ec537b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.562558] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3cafc3-e25c-497a-9a66-f59c382a3727 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.609281] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f9e153-7e5d-4c8e-aa02-f9e976d129c9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.617175] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd68b56-7442-4051-8be2-caaec33e9cf7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.633281] env[63372]: DEBUG nova.compute.provider_tree [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.691621] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024612, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459567} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.691931] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] ddca7aa5-fcc8-4835-bfdb-47781335f2cd/ddca7aa5-fcc8-4835-bfdb-47781335f2cd.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 994.692171] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 994.692431] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8cd2cb83-3f93-4e49-91d9-9d36adb5af33 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.699560] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 994.699560] env[63372]: value = "task-1024613" [ 994.699560] env[63372]: _type = "Task" [ 994.699560] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.709060] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024613, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.726070] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "4a2e53e0-b68c-42be-afff-1eead017679d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.726314] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "4a2e53e0-b68c-42be-afff-1eead017679d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.767711] env[63372]: DEBUG nova.network.neutron [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Successfully updated port: 301c1926-329c-454a-9755-e33bef4550d8 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 994.874920] env[63372]: DEBUG nova.compute.manager [req-702cc327-5477-482f-b0e5-b3d4a04f5b5e req-147835a1-82ca-4f15-be6e-7837d6e37095 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Received event network-vif-plugged-301c1926-329c-454a-9755-e33bef4550d8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.875414] env[63372]: DEBUG oslo_concurrency.lockutils [req-702cc327-5477-482f-b0e5-b3d4a04f5b5e req-147835a1-82ca-4f15-be6e-7837d6e37095 service nova] Acquiring lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.875814] env[63372]: DEBUG oslo_concurrency.lockutils [req-702cc327-5477-482f-b0e5-b3d4a04f5b5e req-147835a1-82ca-4f15-be6e-7837d6e37095 service nova] Lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.876039] env[63372]: DEBUG oslo_concurrency.lockutils [req-702cc327-5477-482f-b0e5-b3d4a04f5b5e req-147835a1-82ca-4f15-be6e-7837d6e37095 service nova] Lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.876221] env[63372]: DEBUG nova.compute.manager [req-702cc327-5477-482f-b0e5-b3d4a04f5b5e req-147835a1-82ca-4f15-be6e-7837d6e37095 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] No waiting events found dispatching network-vif-plugged-301c1926-329c-454a-9755-e33bef4550d8 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 994.876387] env[63372]: WARNING nova.compute.manager [req-702cc327-5477-482f-b0e5-b3d4a04f5b5e req-147835a1-82ca-4f15-be6e-7837d6e37095 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Received unexpected event network-vif-plugged-301c1926-329c-454a-9755-e33bef4550d8 for instance with vm_state building and task_state spawning. [ 994.876657] env[63372]: DEBUG nova.compute.manager [req-702cc327-5477-482f-b0e5-b3d4a04f5b5e req-147835a1-82ca-4f15-be6e-7837d6e37095 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Received event network-changed-301c1926-329c-454a-9755-e33bef4550d8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.876714] env[63372]: DEBUG nova.compute.manager [req-702cc327-5477-482f-b0e5-b3d4a04f5b5e req-147835a1-82ca-4f15-be6e-7837d6e37095 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Refreshing instance network info cache due to event network-changed-301c1926-329c-454a-9755-e33bef4550d8. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 994.876872] env[63372]: DEBUG oslo_concurrency.lockutils [req-702cc327-5477-482f-b0e5-b3d4a04f5b5e req-147835a1-82ca-4f15-be6e-7837d6e37095 service nova] Acquiring lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.877012] env[63372]: DEBUG oslo_concurrency.lockutils [req-702cc327-5477-482f-b0e5-b3d4a04f5b5e req-147835a1-82ca-4f15-be6e-7837d6e37095 service nova] Acquired lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.877194] env[63372]: DEBUG nova.network.neutron [req-702cc327-5477-482f-b0e5-b3d4a04f5b5e req-147835a1-82ca-4f15-be6e-7837d6e37095 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Refreshing network info cache for port 301c1926-329c-454a-9755-e33bef4550d8 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 994.890304] env[63372]: DEBUG nova.network.neutron [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Updating instance_info_cache with network_info: [{"id": "99872bc9-8786-4e6d-9978-42f880ba8407", "address": "fa:16:3e:ab:c5:cb", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99872bc9-87", "ovs_interfaceid": "99872bc9-8786-4e6d-9978-42f880ba8407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.912854] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 994.913919] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b5ed527c-2242-42e8-bb4a-f6539f23caf9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.922029] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 994.922029] env[63372]: value = "task-1024614" [ 994.922029] env[63372]: _type = "Task" [ 994.922029] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.930160] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024614, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.136843] env[63372]: DEBUG nova.scheduler.client.report [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 995.213240] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024613, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065775} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.213551] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 995.214328] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b983a42e-cebb-4473-9dca-67bd45aa5f60 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.230065] env[63372]: DEBUG nova.compute.manager [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 995.241282] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] ddca7aa5-fcc8-4835-bfdb-47781335f2cd/ddca7aa5-fcc8-4835-bfdb-47781335f2cd.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 995.241898] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a1db843-70aa-49ee-a1fa-539a5f64fed2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.262489] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 995.262489] env[63372]: value = "task-1024615" [ 995.262489] env[63372]: _type = "Task" [ 995.262489] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.270717] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024615, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.271916] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.392853] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "refresh_cache-1724ce03-c8e2-415d-a380-59ac69fbfb57" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.393623] env[63372]: DEBUG nova.compute.manager [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Instance network_info: |[{"id": "99872bc9-8786-4e6d-9978-42f880ba8407", "address": "fa:16:3e:ab:c5:cb", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99872bc9-87", "ovs_interfaceid": "99872bc9-8786-4e6d-9978-42f880ba8407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 995.394181] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ab:c5:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9f208df-1fb5-4403-9796-7fd19e4bfb85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '99872bc9-8786-4e6d-9978-42f880ba8407', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 995.402282] env[63372]: DEBUG oslo.service.loopingcall [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.402891] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 995.403748] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6774a3d2-4669-41ec-af80-883d211110f6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.426865] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 995.426865] env[63372]: value = "task-1024616" [ 995.426865] env[63372]: _type = "Task" [ 995.426865] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.435270] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024614, 'name': CreateSnapshot_Task, 'duration_secs': 0.442355} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.435270] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 995.435270] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a376448-c96e-4df1-8bfd-6d976b0d1fab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.440428] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024616, 'name': CreateVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.441084] env[63372]: DEBUG nova.network.neutron [req-702cc327-5477-482f-b0e5-b3d4a04f5b5e req-147835a1-82ca-4f15-be6e-7837d6e37095 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 995.589040] env[63372]: DEBUG nova.network.neutron [req-702cc327-5477-482f-b0e5-b3d4a04f5b5e req-147835a1-82ca-4f15-be6e-7837d6e37095 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.641190] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.334s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.641426] env[63372]: INFO nova.compute.manager [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Migrating [ 995.647647] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.384s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.647824] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.647974] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63372) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 995.648282] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.930s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.648497] env[63372]: DEBUG nova.objects.instance [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lazy-loading 'resources' on Instance uuid a3145c98-e046-4b55-b80a-57e63807bd7a {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.652288] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd12cb25-32be-4815-9924-38ec84e1a58a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.668899] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087df11a-2643-4498-a506-f8aabd93de7d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.686904] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdbe2c6e-2a9a-47c8-bd24-dac4e0af1d27 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.694298] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b132ab63-b24c-4b0e-ab2a-afedf56b411e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.726857] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180938MB free_disk=185GB free_vcpus=48 pci_devices=None {{(pid=63372) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 995.727032] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.755267] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.771871] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024615, 'name': ReconfigVM_Task, 'duration_secs': 0.274451} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.774753] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Reconfigured VM instance instance-00000060 to attach disk [datastore2] ddca7aa5-fcc8-4835-bfdb-47781335f2cd/ddca7aa5-fcc8-4835-bfdb-47781335f2cd.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 995.776108] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0661e8dc-39dd-4140-8842-dc80120f01e0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.781661] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 995.781661] env[63372]: value = "task-1024617" [ 995.781661] env[63372]: _type = "Task" [ 995.781661] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.793068] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024617, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.872323] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf3fbc9-b588-4e33-ab03-332f5bf3f5fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.880270] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c391d4a2-3826-4299-a987-09c82c4b604c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.910863] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82684191-46bf-4f8b-9c40-ca13d583346d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.918771] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b2a320-1eb6-4a24-ab00-64877ed38dbe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.934449] env[63372]: DEBUG nova.compute.provider_tree [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 995.944872] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024616, 'name': CreateVM_Task, 'duration_secs': 0.403004} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.945069] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 995.945670] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.945808] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.946580] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 995.947098] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39cecf2f-bfb0-4999-b038-edfd319c0ec8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.959661] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 995.961453] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c935aa60-af69-4355-b980-20de99c8f90d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.967492] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 995.967492] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525723a0-cdf0-a13b-2b29-8dcf896471d2" [ 995.967492] env[63372]: _type = "Task" [ 995.967492] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.970675] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 995.970675] env[63372]: value = "task-1024618" [ 995.970675] env[63372]: _type = "Task" [ 995.970675] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.974691] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525723a0-cdf0-a13b-2b29-8dcf896471d2, 'name': SearchDatastore_Task, 'duration_secs': 0.009352} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.977660] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.978188] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.978188] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.978307] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.978481] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.978905] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f93a96a-ebc5-4a85-8fa4-a788bbeb172b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.985853] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024618, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.989615] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.989783] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 995.990491] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86f02fbd-fb37-4f33-80fa-67d183c6cdff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.997234] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 995.997234] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]522b7def-e434-0956-db5c-7fbd7c62be6d" [ 995.997234] env[63372]: _type = "Task" [ 995.997234] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.002486] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522b7def-e434-0956-db5c-7fbd7c62be6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.091556] env[63372]: DEBUG oslo_concurrency.lockutils [req-702cc327-5477-482f-b0e5-b3d4a04f5b5e req-147835a1-82ca-4f15-be6e-7837d6e37095 service nova] Releasing lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.091891] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.092070] env[63372]: DEBUG nova.network.neutron [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 996.163857] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.163990] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.164166] env[63372]: DEBUG nova.network.neutron [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 996.291781] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024617, 'name': Rename_Task, 'duration_secs': 0.181488} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.292105] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 996.292365] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97b5cada-6bbd-4beb-9fb5-76fd23a8f175 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.298799] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 996.298799] env[63372]: value = "task-1024619" [ 996.298799] env[63372]: _type = "Task" [ 996.298799] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.306860] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024619, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.460189] env[63372]: ERROR nova.scheduler.client.report [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [req-a24b9721-cfe8-472a-af5d-52f1055aa1f3] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a24b9721-cfe8-472a-af5d-52f1055aa1f3"}]} [ 996.480399] env[63372]: DEBUG nova.scheduler.client.report [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Refreshing inventories for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 996.485927] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "28eb1675-b1b8-46a6-873d-5c858b716575" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.487273] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "28eb1675-b1b8-46a6-873d-5c858b716575" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.497935] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024618, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.508775] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522b7def-e434-0956-db5c-7fbd7c62be6d, 'name': SearchDatastore_Task, 'duration_secs': 0.008156} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.509791] env[63372]: DEBUG nova.scheduler.client.report [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Updating ProviderTree inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 996.510041] env[63372]: DEBUG nova.compute.provider_tree [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 996.513222] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20200954-b134-4ed7-b4a6-1b5fa6807624 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.518874] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 996.518874] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52d9fc56-a0e8-aedf-13d9-de178ea71a96" [ 996.518874] env[63372]: _type = "Task" [ 996.518874] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.523581] env[63372]: DEBUG nova.scheduler.client.report [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Refreshing aggregate associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, aggregates: None {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 996.528663] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52d9fc56-a0e8-aedf-13d9-de178ea71a96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.545433] env[63372]: DEBUG nova.scheduler.client.report [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Refreshing trait associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 996.625703] env[63372]: DEBUG nova.network.neutron [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 996.720942] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96757953-36e4-41d7-b102-e252f8f97540 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.728858] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef7b3c7-db9e-4280-891d-bcff1f858327 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.764483] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec9c18e-c505-44a3-b852-c4b7a11c9c5b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.772949] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397ce298-a212-413d-84dd-22d908e56899 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.787680] env[63372]: DEBUG nova.compute.provider_tree [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 996.808279] env[63372]: DEBUG oslo_vmware.api [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024619, 'name': PowerOnVM_Task, 'duration_secs': 0.446664} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.808590] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 996.808812] env[63372]: INFO nova.compute.manager [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Took 7.88 seconds to spawn the instance on the hypervisor. [ 996.809057] env[63372]: DEBUG nova.compute.manager [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 996.809808] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548ce64f-5c0c-4fc8-aa9c-fae40f8751f4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.918816] env[63372]: DEBUG nova.network.neutron [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Updating instance_info_cache with network_info: [{"id": "301c1926-329c-454a-9755-e33bef4550d8", "address": "fa:16:3e:bc:d9:91", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap301c1926-32", "ovs_interfaceid": "301c1926-329c-454a-9755-e33bef4550d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.987504] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024618, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.002249] env[63372]: DEBUG nova.compute.manager [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 997.036862] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52d9fc56-a0e8-aedf-13d9-de178ea71a96, 'name': SearchDatastore_Task, 'duration_secs': 0.008758} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.037371] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.037809] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 1724ce03-c8e2-415d-a380-59ac69fbfb57/1724ce03-c8e2-415d-a380-59ac69fbfb57.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 997.038195] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c3e13e4-3123-4e40-b793-bf37c6892fbf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.046556] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 997.046556] env[63372]: value = "task-1024620" [ 997.046556] env[63372]: _type = "Task" [ 997.046556] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.057790] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024620, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.068271] env[63372]: DEBUG nova.network.neutron [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating instance_info_cache with network_info: [{"id": "298d217f-c17a-4ee2-acae-baee355112dd", "address": "fa:16:3e:66:af:86", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298d217f-c1", "ovs_interfaceid": "298d217f-c17a-4ee2-acae-baee355112dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.319878] env[63372]: DEBUG nova.scheduler.client.report [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 137 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 997.320222] env[63372]: DEBUG nova.compute.provider_tree [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 137 to 138 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 997.320465] env[63372]: DEBUG nova.compute.provider_tree [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 997.332025] env[63372]: INFO nova.compute.manager [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Took 20.95 seconds to build instance. [ 997.422143] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.422559] env[63372]: DEBUG nova.compute.manager [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Instance network_info: |[{"id": "301c1926-329c-454a-9755-e33bef4550d8", "address": "fa:16:3e:bc:d9:91", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap301c1926-32", "ovs_interfaceid": "301c1926-329c-454a-9755-e33bef4550d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 997.423039] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:d9:91', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '301c1926-329c-454a-9755-e33bef4550d8', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 997.432143] env[63372]: DEBUG oslo.service.loopingcall [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.432460] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 997.432794] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-13a53d00-eb80-4989-a264-8970f102b637 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.457381] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 997.457381] env[63372]: value = "task-1024621" [ 997.457381] env[63372]: _type = "Task" [ 997.457381] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.467828] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024621, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.485660] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024618, 'name': CloneVM_Task, 'duration_secs': 1.257088} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.487026] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Created linked-clone VM from snapshot [ 997.487026] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3929b96-6407-477b-8d4b-b5e6d97aac42 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.494437] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Uploading image 243dbf27-41ba-48e9-ac46-97223c25bb0e {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 997.522775] env[63372]: DEBUG oslo_vmware.rw_handles [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 997.522775] env[63372]: value = "vm-227440" [ 997.522775] env[63372]: _type = "VirtualMachine" [ 997.522775] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 997.523132] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-779f73d7-35a4-4444-8008-82c0931a7e26 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.525541] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.531047] env[63372]: DEBUG oslo_vmware.rw_handles [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lease: (returnval){ [ 997.531047] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52690edf-5a35-a99b-3e7a-ebd09bbcc1fb" [ 997.531047] env[63372]: _type = "HttpNfcLease" [ 997.531047] env[63372]: } obtained for exporting VM: (result){ [ 997.531047] env[63372]: value = "vm-227440" [ 997.531047] env[63372]: _type = "VirtualMachine" [ 997.531047] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 997.531442] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the lease: (returnval){ [ 997.531442] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52690edf-5a35-a99b-3e7a-ebd09bbcc1fb" [ 997.531442] env[63372]: _type = "HttpNfcLease" [ 997.531442] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 997.538149] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 997.538149] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52690edf-5a35-a99b-3e7a-ebd09bbcc1fb" [ 997.538149] env[63372]: _type = "HttpNfcLease" [ 997.538149] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 997.556385] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024620, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461924} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.556673] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 1724ce03-c8e2-415d-a380-59ac69fbfb57/1724ce03-c8e2-415d-a380-59ac69fbfb57.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 997.556910] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.557182] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d9033a13-e867-4737-a508-6279c684e047 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.563544] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 997.563544] env[63372]: value = "task-1024623" [ 997.563544] env[63372]: _type = "Task" [ 997.563544] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.571981] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.574161] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024623, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.834403] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d710b13-5532-4276-9ce8-8108e5fe32fc tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.471s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.835194] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.187s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.837588] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.111s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.857361] env[63372]: INFO nova.scheduler.client.report [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleted allocations for instance a3145c98-e046-4b55-b80a-57e63807bd7a [ 997.966930] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024621, 'name': CreateVM_Task, 'duration_secs': 0.340617} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.967184] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 997.967889] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.968078] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.968403] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 997.968991] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0708a4ff-2b24-442f-85bb-71e5b8151c58 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.974078] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 997.974078] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52f19a4e-396d-2833-afe7-9dac5a9793a6" [ 997.974078] env[63372]: _type = "Task" [ 997.974078] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.981678] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f19a4e-396d-2833-afe7-9dac5a9793a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.039926] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 998.039926] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52690edf-5a35-a99b-3e7a-ebd09bbcc1fb" [ 998.039926] env[63372]: _type = "HttpNfcLease" [ 998.039926] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 998.040324] env[63372]: DEBUG oslo_vmware.rw_handles [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 998.040324] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52690edf-5a35-a99b-3e7a-ebd09bbcc1fb" [ 998.040324] env[63372]: _type = "HttpNfcLease" [ 998.040324] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 998.041148] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fdea35d-bf8e-409f-b7cb-58acedf92c84 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.050175] env[63372]: DEBUG oslo_vmware.rw_handles [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52013295-109e-267e-17d8-e42f47fad683/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 998.050353] env[63372]: DEBUG oslo_vmware.rw_handles [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52013295-109e-267e-17d8-e42f47fad683/disk-0.vmdk for reading. {{(pid=63372) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 998.120672] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024623, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.054662} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.120985] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 998.121874] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b75a3f4-8c5a-4b9b-9015-2f709081d704 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.145208] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 1724ce03-c8e2-415d-a380-59ac69fbfb57/1724ce03-c8e2-415d-a380-59ac69fbfb57.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 998.147764] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4dc31d9e-f077-429c-9bde-af64867c0ebe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.163673] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5426daff-4f4a-4901-a413-a7da8a22e284 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.170918] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 998.170918] env[63372]: value = "task-1024624" [ 998.170918] env[63372]: _type = "Task" [ 998.170918] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.179526] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024624, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.366641] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6072add-2b04-458e-be4b-704bd5830abf tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "a3145c98-e046-4b55-b80a-57e63807bd7a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.064s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.484744] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52f19a4e-396d-2833-afe7-9dac5a9793a6, 'name': SearchDatastore_Task, 'duration_secs': 0.00884} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.485130] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.485561] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 998.485841] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.486061] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.486409] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 998.486871] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d29a6af9-365a-4b0c-906b-d99387f206fd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.495981] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 998.496206] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 998.497192] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d59041f-1ebd-4cf5-8adb-c33930c1557b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.502953] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 998.502953] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52834526-7772-bbfd-c572-09b3f515961b" [ 998.502953] env[63372]: _type = "Task" [ 998.502953] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.514113] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52834526-7772-bbfd-c572-09b3f515961b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.683131] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024624, 'name': ReconfigVM_Task, 'duration_secs': 0.295468} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.684109] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 1724ce03-c8e2-415d-a380-59ac69fbfb57/1724ce03-c8e2-415d-a380-59ac69fbfb57.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.684812] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ecbb984a-6035-47bf-a6ee-dac2b6eeb9ae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.695760] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 998.695760] env[63372]: value = "task-1024625" [ 998.695760] env[63372]: _type = "Task" [ 998.695760] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.706745] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024625, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.851059] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Applying migration context for instance cbb471c5-8d86-4d8f-8935-f6a77e5144be as it has an incoming, in-progress migration 75fa60a2-d2ae-4672-8d3c-8a09a4cdd49a. Migration status is migrating {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 998.852570] env[63372]: INFO nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating resource usage from migration 75fa60a2-d2ae-4672-8d3c-8a09a4cdd49a [ 998.885499] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance c4718797-aa86-4ec0-94d3-6480bd6aa898 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 998.885773] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 91985614-b959-401e-bb06-d67b230ee026 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 998.885841] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance b7732621-7e58-40a3-b723-5c66df6f74a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 998.885981] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance b455b16b-3332-43bf-a91e-a350287ba5f3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 998.886135] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 9be70e66-2d81-4c7b-8dcc-e69815faef1a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 998.886261] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance ddca7aa5-fcc8-4835-bfdb-47781335f2cd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 998.886490] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 1724ce03-c8e2-415d-a380-59ac69fbfb57 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 998.886604] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance d7b7ee35-5e83-4c62-bd1e-8ec39951c44c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 998.886715] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Migration 75fa60a2-d2ae-4672-8d3c-8a09a4cdd49a is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 998.886822] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance cbb471c5-8d86-4d8f-8935-f6a77e5144be actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 999.022316] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52834526-7772-bbfd-c572-09b3f515961b, 'name': SearchDatastore_Task, 'duration_secs': 0.023738} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.023603] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6cb0dd3-4c92-4ad4-aeec-97b6c44f9197 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.030311] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 999.030311] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52a8847b-bbbf-d48b-23c6-2810b9bde7b4" [ 999.030311] env[63372]: _type = "Task" [ 999.030311] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.043752] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a8847b-bbbf-d48b-23c6-2810b9bde7b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.206736] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024625, 'name': Rename_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.340371] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-810408ab-ad78-42ba-b61c-e00cad81b47b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.372676] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating instance 'cbb471c5-8d86-4d8f-8935-f6a77e5144be' progress to 0 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 999.390927] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 4a2e53e0-b68c-42be-afff-1eead017679d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 999.544553] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a8847b-bbbf-d48b-23c6-2810b9bde7b4, 'name': SearchDatastore_Task, 'duration_secs': 0.017329} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.544771] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.545246] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] d7b7ee35-5e83-4c62-bd1e-8ec39951c44c/d7b7ee35-5e83-4c62-bd1e-8ec39951c44c.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 999.545606] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88fcd107-1d98-48db-94c6-946eab5bfe53 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.553018] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 999.553018] env[63372]: value = "task-1024626" [ 999.553018] env[63372]: _type = "Task" [ 999.553018] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.562145] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024626, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.707355] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024625, 'name': Rename_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.880149] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 999.880567] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60e4d48c-1558-484a-9036-6728953bbda4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.887809] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 999.887809] env[63372]: value = "task-1024627" [ 999.887809] env[63372]: _type = "Task" [ 999.887809] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.897032] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 28eb1675-b1b8-46a6-873d-5c858b716575 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 999.897200] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 999.897398] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 999.899772] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024627, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.066505] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.066505] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.081706] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024626, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.113700] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f760b82-1e39-45e5-9a59-dcc9a45bda26 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.122960] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627a95d0-f188-44d0-915a-3b45dcca350f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.156025] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6dd224e-0845-4156-bd94-980125c67a9d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.164399] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffefa3e9-777a-48d8-bf33-92b5e8939679 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.179420] env[63372]: DEBUG nova.compute.provider_tree [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.207411] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024625, 'name': Rename_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.398350] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024627, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.567932] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024626, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.574091] env[63372]: DEBUG nova.compute.manager [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1000.682992] env[63372]: DEBUG nova.scheduler.client.report [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.709183] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024625, 'name': Rename_Task, 'duration_secs': 1.569358} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.709585] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1000.709913] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3554bd60-7ec3-454f-a323-31736aecba69 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.715940] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1000.715940] env[63372]: value = "task-1024628" [ 1000.715940] env[63372]: _type = "Task" [ 1000.715940] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.724473] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024628, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.901545] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024627, 'name': PowerOffVM_Task, 'duration_secs': 0.562608} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.901785] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1000.901998] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating instance 'cbb471c5-8d86-4d8f-8935-f6a77e5144be' progress to 17 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1001.066792] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024626, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.173394} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.067113] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] d7b7ee35-5e83-4c62-bd1e-8ec39951c44c/d7b7ee35-5e83-4c62-bd1e-8ec39951c44c.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1001.067354] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1001.067622] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b1c3119-b121-46be-bacb-24d3893f415f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.074694] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 1001.074694] env[63372]: value = "task-1024629" [ 1001.074694] env[63372]: _type = "Task" [ 1001.074694] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.085952] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024629, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.099672] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.188961] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1001.189235] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.352s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.189561] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.434s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.191314] env[63372]: INFO nova.compute.claims [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.227297] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024628, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.409014] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1001.409549] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1001.409792] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1001.410111] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1001.410363] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1001.410633] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1001.410944] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1001.411184] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1001.411372] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1001.411590] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1001.411737] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1001.417169] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-584d94ac-c27d-4e4d-87d7-637e99a1aebb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.433685] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1001.433685] env[63372]: value = "task-1024630" [ 1001.433685] env[63372]: _type = "Task" [ 1001.433685] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.442257] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024630, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.584757] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024629, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081203} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.585053] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1001.585834] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daaeee6a-3563-4f5d-ab43-0ba1fa4c6a92 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.607896] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] d7b7ee35-5e83-4c62-bd1e-8ec39951c44c/d7b7ee35-5e83-4c62-bd1e-8ec39951c44c.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.608333] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77e8ca57-4c74-4847-8b23-c642a97b4733 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.628182] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 1001.628182] env[63372]: value = "task-1024631" [ 1001.628182] env[63372]: _type = "Task" [ 1001.628182] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.636022] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024631, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.726410] env[63372]: DEBUG oslo_vmware.api [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024628, 'name': PowerOnVM_Task, 'duration_secs': 0.739363} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.726687] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1001.726891] env[63372]: INFO nova.compute.manager [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Took 10.50 seconds to spawn the instance on the hypervisor. [ 1001.727077] env[63372]: DEBUG nova.compute.manager [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.727872] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c29e43-f4b3-460a-8842-a1b666c8de36 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.734060] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "91985614-b959-401e-bb06-d67b230ee026" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.734378] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "91985614-b959-401e-bb06-d67b230ee026" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.734577] env[63372]: INFO nova.compute.manager [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Shelving [ 1001.945036] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024630, 'name': ReconfigVM_Task, 'duration_secs': 0.236667} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.945036] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating instance 'cbb471c5-8d86-4d8f-8935-f6a77e5144be' progress to 33 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1002.140350] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024631, 'name': ReconfigVM_Task, 'duration_secs': 0.505122} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.140350] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Reconfigured VM instance instance-00000062 to attach disk [datastore2] d7b7ee35-5e83-4c62-bd1e-8ec39951c44c/d7b7ee35-5e83-4c62-bd1e-8ec39951c44c.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1002.140516] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eedd102e-15af-49d6-af2e-725f5f46caf4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.146467] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 1002.146467] env[63372]: value = "task-1024632" [ 1002.146467] env[63372]: _type = "Task" [ 1002.146467] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.154883] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024632, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.250111] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1002.250111] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4aeb293-ba71-440d-9483-6989da7cc75f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.251969] env[63372]: INFO nova.compute.manager [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Took 25.44 seconds to build instance. [ 1002.260953] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1002.260953] env[63372]: value = "task-1024633" [ 1002.260953] env[63372]: _type = "Task" [ 1002.260953] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.272260] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024633, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.391489] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2162d63a-1330-49e6-93b9-4a5ebb98a192 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.400569] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8eac8e-769d-4e2f-a2ec-3307c837c344 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.432963] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da7ab71-e7e2-49f7-a70f-dd8cfd502f27 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.440295] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4580a56f-00c4-4a5a-901f-485d64de5d95 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.456569] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1002.456795] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1002.456843] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1002.458026] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1002.458026] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1002.458026] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1002.458026] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1002.458026] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1002.458026] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1002.458326] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1002.458406] env[63372]: DEBUG nova.virt.hardware [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1002.464053] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Reconfiguring VM instance instance-0000005d to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1002.464630] env[63372]: DEBUG nova.compute.provider_tree [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.466514] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d95faa8-880e-4a83-8b40-f8b8d3151b79 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.486545] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1002.486545] env[63372]: value = "task-1024634" [ 1002.486545] env[63372]: _type = "Task" [ 1002.486545] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.494914] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024634, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.657756] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024632, 'name': Rename_Task, 'duration_secs': 0.19393} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.658069] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1002.658595] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c3b8486b-2e09-490e-893c-a05eca21a5d3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.665045] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 1002.665045] env[63372]: value = "task-1024635" [ 1002.665045] env[63372]: _type = "Task" [ 1002.665045] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.673241] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024635, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.715075] env[63372]: INFO nova.compute.manager [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Rescuing [ 1002.715410] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "refresh_cache-1724ce03-c8e2-415d-a380-59ac69fbfb57" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.715574] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "refresh_cache-1724ce03-c8e2-415d-a380-59ac69fbfb57" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.715888] env[63372]: DEBUG nova.network.neutron [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1002.756489] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f3612436-cd06-4374-9a03-4ff2177ea430 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "1724ce03-c8e2-415d-a380-59ac69fbfb57" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.957s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.770984] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024633, 'name': PowerOffVM_Task, 'duration_secs': 0.243623} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.771376] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1002.772273] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3756eaf5-d2a7-4514-b2bd-2b13809f1a9d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.790546] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744181ad-bb4e-4467-b92f-05196013426b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.968063] env[63372]: DEBUG nova.scheduler.client.report [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1002.997825] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024634, 'name': ReconfigVM_Task, 'duration_secs': 0.24452} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.998131] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Reconfigured VM instance instance-0000005d to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1002.998928] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049c5218-cdd1-48d0-917f-7420a1bd2633 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.024656] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] cbb471c5-8d86-4d8f-8935-f6a77e5144be/cbb471c5-8d86-4d8f-8935-f6a77e5144be.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1003.024975] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12f0683f-f6fc-4491-b9e2-0be5f6f7178f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.045143] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1003.045143] env[63372]: value = "task-1024636" [ 1003.045143] env[63372]: _type = "Task" [ 1003.045143] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.054215] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024636, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.175766] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024635, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.301068] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1003.301421] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-db80b743-a484-4dc5-b998-7381e808d042 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.308896] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1003.308896] env[63372]: value = "task-1024637" [ 1003.308896] env[63372]: _type = "Task" [ 1003.308896] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.318444] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024637, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.472745] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.473536] env[63372]: DEBUG nova.compute.manager [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1003.476890] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.951s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.478807] env[63372]: INFO nova.compute.claims [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1003.556347] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024636, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.570924] env[63372]: DEBUG nova.network.neutron [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Updating instance_info_cache with network_info: [{"id": "99872bc9-8786-4e6d-9978-42f880ba8407", "address": "fa:16:3e:ab:c5:cb", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99872bc9-87", "ovs_interfaceid": "99872bc9-8786-4e6d-9978-42f880ba8407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.679375] env[63372]: DEBUG oslo_vmware.api [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024635, 'name': PowerOnVM_Task, 'duration_secs': 0.766438} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.679799] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1003.680136] env[63372]: INFO nova.compute.manager [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Took 10.13 seconds to spawn the instance on the hypervisor. [ 1003.680429] env[63372]: DEBUG nova.compute.manager [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1003.681738] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4a7954-4705-4083-99c8-39d11ef49a0f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.818962] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024637, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.978838] env[63372]: DEBUG nova.compute.utils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1003.980811] env[63372]: DEBUG nova.compute.manager [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1003.981075] env[63372]: DEBUG nova.network.neutron [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1004.028062] env[63372]: DEBUG nova.policy [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d822682f06640489a48f8577acf4389', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e46b6acae4a4b71bc390bd513ddd2f1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1004.057883] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024636, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.074300] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "refresh_cache-1724ce03-c8e2-415d-a380-59ac69fbfb57" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.205656] env[63372]: INFO nova.compute.manager [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Took 25.90 seconds to build instance. [ 1004.321691] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024637, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.484088] env[63372]: DEBUG nova.compute.manager [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1004.507502] env[63372]: DEBUG nova.network.neutron [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Successfully created port: 7cd17ebc-3698-499b-8a5a-0507f100485c {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1004.558282] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024636, 'name': ReconfigVM_Task, 'duration_secs': 1.210737} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.558569] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Reconfigured VM instance instance-0000005d to attach disk [datastore1] cbb471c5-8d86-4d8f-8935-f6a77e5144be/cbb471c5-8d86-4d8f-8935-f6a77e5144be.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1004.558861] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating instance 'cbb471c5-8d86-4d8f-8935-f6a77e5144be' progress to 50 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1004.612840] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1004.613228] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a4353c8-bb91-4849-82f9-401cb5934f4e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.620573] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1004.620573] env[63372]: value = "task-1024638" [ 1004.620573] env[63372]: _type = "Task" [ 1004.620573] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.635443] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024638, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.708957] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d57fb9e7-b67b-48b5-af1c-10c1b09bbbc6 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.409s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.756756] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3b7dd7-fbeb-4ad6-aebb-9a5af291ccea {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.765564] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70497af9-1e73-4f63-807c-5f4af8a3bfe2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.797680] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35434f9d-d37c-4554-a7fe-12ddd4c98820 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.805744] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f678f4ff-233f-4518-baee-e41264663a54 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.824027] env[63372]: DEBUG nova.compute.provider_tree [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.830904] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024637, 'name': CreateSnapshot_Task, 'duration_secs': 1.103608} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.831274] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1004.832113] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8601823-75a0-4916-8f17-7b01d8d83c73 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.068350] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-271c47ac-7955-4c6c-97c6-360fb1e2c12d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.090217] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63b0e9f-978b-4107-8889-28b0a1721ae4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.112027] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating instance 'cbb471c5-8d86-4d8f-8935-f6a77e5144be' progress to 67 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1005.133170] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024638, 'name': PowerOffVM_Task, 'duration_secs': 0.210133} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.133762] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1005.134502] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910c1a10-1e80-4b2e-88fb-27454e5dce44 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.153925] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6374525b-14de-461f-b0c7-11b71ca3cdeb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.195615] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1005.195963] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b214e947-548e-4abd-b7c2-251b25cdaedb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.204050] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1005.204050] env[63372]: value = "task-1024639" [ 1005.204050] env[63372]: _type = "Task" [ 1005.204050] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.214752] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] VM already powered off {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1005.215014] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1005.215295] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.215474] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.215674] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1005.215968] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2521035-5a7f-4079-a244-206d63fb226e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.225486] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1005.225901] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1005.226643] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1deeeec8-0d27-444a-b8d6-6adafd7298e5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.238254] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1005.238254] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b8d480-b966-5e80-b4da-cff6ccf7c62f" [ 1005.238254] env[63372]: _type = "Task" [ 1005.238254] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.246767] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b8d480-b966-5e80-b4da-cff6ccf7c62f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.330019] env[63372]: DEBUG nova.scheduler.client.report [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1005.335171] env[63372]: DEBUG nova.compute.manager [req-815f0625-bd7e-4457-a930-850636f22432 req-afaaa458-676a-4e47-a850-1715da802f78 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Received event network-changed-18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.335171] env[63372]: DEBUG nova.compute.manager [req-815f0625-bd7e-4457-a930-850636f22432 req-afaaa458-676a-4e47-a850-1715da802f78 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Refreshing instance network info cache due to event network-changed-18c7850a-d97e-4564-92e5-38d1204369e4. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1005.335171] env[63372]: DEBUG oslo_concurrency.lockutils [req-815f0625-bd7e-4457-a930-850636f22432 req-afaaa458-676a-4e47-a850-1715da802f78 service nova] Acquiring lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.335171] env[63372]: DEBUG oslo_concurrency.lockutils [req-815f0625-bd7e-4457-a930-850636f22432 req-afaaa458-676a-4e47-a850-1715da802f78 service nova] Acquired lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.335418] env[63372]: DEBUG nova.network.neutron [req-815f0625-bd7e-4457-a930-850636f22432 req-afaaa458-676a-4e47-a850-1715da802f78 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Refreshing network info cache for port 18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1005.357331] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1005.358878] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-166c1638-d25c-4ba8-a115-f9fe6cde5c44 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.372181] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1005.372181] env[63372]: value = "task-1024640" [ 1005.372181] env[63372]: _type = "Task" [ 1005.372181] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.383864] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024640, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.503661] env[63372]: DEBUG nova.compute.manager [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1005.536675] env[63372]: DEBUG nova.virt.hardware [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.537767] env[63372]: DEBUG nova.virt.hardware [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.537767] env[63372]: DEBUG nova.virt.hardware [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.537767] env[63372]: DEBUG nova.virt.hardware [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.537767] env[63372]: DEBUG nova.virt.hardware [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.537972] env[63372]: DEBUG nova.virt.hardware [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.538335] env[63372]: DEBUG nova.virt.hardware [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.538551] env[63372]: DEBUG nova.virt.hardware [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.538742] env[63372]: DEBUG nova.virt.hardware [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.539505] env[63372]: DEBUG nova.virt.hardware [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.539505] env[63372]: DEBUG nova.virt.hardware [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.540024] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c01673-afbf-452c-80ba-10777e8a6389 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.548820] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6929c3fe-e1b1-4c20-bf9a-0e9c9872c5ce {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.598227] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "b7732621-7e58-40a3-b723-5c66df6f74a1" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.598568] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "b7732621-7e58-40a3-b723-5c66df6f74a1" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.678375] env[63372]: DEBUG nova.network.neutron [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Port 298d217f-c17a-4ee2-acae-baee355112dd binding to destination host cpu-1 is already ACTIVE {{(pid=63372) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1005.753630] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b8d480-b966-5e80-b4da-cff6ccf7c62f, 'name': SearchDatastore_Task, 'duration_secs': 0.009609} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.754783] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b72ba7de-59b7-49de-8982-1e1dadfcfdcf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.760589] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1005.760589] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52862647-c25f-9e90-6525-93a638eac86d" [ 1005.760589] env[63372]: _type = "Task" [ 1005.760589] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.769754] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52862647-c25f-9e90-6525-93a638eac86d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.838375] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.361s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.838955] env[63372]: DEBUG nova.compute.manager [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1005.844161] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.745s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.845718] env[63372]: INFO nova.compute.claims [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.883253] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024640, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.102717] env[63372]: DEBUG nova.compute.utils [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1006.223563] env[63372]: DEBUG nova.network.neutron [req-815f0625-bd7e-4457-a930-850636f22432 req-afaaa458-676a-4e47-a850-1715da802f78 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updated VIF entry in instance network info cache for port 18c7850a-d97e-4564-92e5-38d1204369e4. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1006.223962] env[63372]: DEBUG nova.network.neutron [req-815f0625-bd7e-4457-a930-850636f22432 req-afaaa458-676a-4e47-a850-1715da802f78 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updating instance_info_cache with network_info: [{"id": "18c7850a-d97e-4564-92e5-38d1204369e4", "address": "fa:16:3e:14:28:4a", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c7850a-d9", "ovs_interfaceid": "18c7850a-d97e-4564-92e5-38d1204369e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.271189] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52862647-c25f-9e90-6525-93a638eac86d, 'name': SearchDatastore_Task, 'duration_secs': 0.019581} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.271462] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.272636] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 1724ce03-c8e2-415d-a380-59ac69fbfb57/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk. {{(pid=63372) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1006.272950] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5135299-4148-4665-b12a-b1204040e49e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.280143] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1006.280143] env[63372]: value = "task-1024641" [ 1006.280143] env[63372]: _type = "Task" [ 1006.280143] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.288702] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024641, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.350435] env[63372]: DEBUG nova.compute.utils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1006.354407] env[63372]: DEBUG nova.compute.manager [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1006.354606] env[63372]: DEBUG nova.network.neutron [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1006.384721] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024640, 'name': CloneVM_Task} progress is 95%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.454834] env[63372]: DEBUG nova.policy [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ddaa4503a23947a9a06962ed4d01aa32', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3d866ac9602470ea9f8d8159551acc8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1006.606563] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "b7732621-7e58-40a3-b723-5c66df6f74a1" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.627518] env[63372]: DEBUG nova.network.neutron [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Successfully updated port: 7cd17ebc-3698-499b-8a5a-0507f100485c {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1006.707880] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.708229] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.708329] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.724185] env[63372]: DEBUG nova.compute.manager [req-cbe18190-2ee2-4bbc-9b66-c783b652429c req-801cb891-71e2-4b95-a0db-58a6e6f10323 service nova] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Received event network-vif-plugged-7cd17ebc-3698-499b-8a5a-0507f100485c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.724185] env[63372]: DEBUG oslo_concurrency.lockutils [req-cbe18190-2ee2-4bbc-9b66-c783b652429c req-801cb891-71e2-4b95-a0db-58a6e6f10323 service nova] Acquiring lock "4a2e53e0-b68c-42be-afff-1eead017679d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.724185] env[63372]: DEBUG oslo_concurrency.lockutils [req-cbe18190-2ee2-4bbc-9b66-c783b652429c req-801cb891-71e2-4b95-a0db-58a6e6f10323 service nova] Lock "4a2e53e0-b68c-42be-afff-1eead017679d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.724862] env[63372]: DEBUG oslo_concurrency.lockutils [req-cbe18190-2ee2-4bbc-9b66-c783b652429c req-801cb891-71e2-4b95-a0db-58a6e6f10323 service nova] Lock "4a2e53e0-b68c-42be-afff-1eead017679d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.725292] env[63372]: DEBUG nova.compute.manager [req-cbe18190-2ee2-4bbc-9b66-c783b652429c req-801cb891-71e2-4b95-a0db-58a6e6f10323 service nova] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] No waiting events found dispatching network-vif-plugged-7cd17ebc-3698-499b-8a5a-0507f100485c {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1006.725644] env[63372]: WARNING nova.compute.manager [req-cbe18190-2ee2-4bbc-9b66-c783b652429c req-801cb891-71e2-4b95-a0db-58a6e6f10323 service nova] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Received unexpected event network-vif-plugged-7cd17ebc-3698-499b-8a5a-0507f100485c for instance with vm_state building and task_state spawning. [ 1006.728112] env[63372]: DEBUG oslo_concurrency.lockutils [req-815f0625-bd7e-4457-a930-850636f22432 req-afaaa458-676a-4e47-a850-1715da802f78 service nova] Releasing lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.798455] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024641, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.855711] env[63372]: DEBUG nova.compute.manager [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1006.893432] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024640, 'name': CloneVM_Task, 'duration_secs': 1.470559} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.893887] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Created linked-clone VM from snapshot [ 1006.897023] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c46a445-e7b1-4fc8-b87d-4cc0fcb46e0b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.903331] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Uploading image 003d2e97-266b-4459-abd7-6ca82ed4825e {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1006.931360] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1006.931360] env[63372]: value = "vm-227443" [ 1006.931360] env[63372]: _type = "VirtualMachine" [ 1006.931360] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1006.931952] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-402dfb4e-0d53-45f9-9e61-3f3f56af5c31 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.939096] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lease: (returnval){ [ 1006.939096] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52623459-9c36-e714-df7d-df9c4b8282ea" [ 1006.939096] env[63372]: _type = "HttpNfcLease" [ 1006.939096] env[63372]: } obtained for exporting VM: (result){ [ 1006.939096] env[63372]: value = "vm-227443" [ 1006.939096] env[63372]: _type = "VirtualMachine" [ 1006.939096] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1006.939424] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the lease: (returnval){ [ 1006.939424] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52623459-9c36-e714-df7d-df9c4b8282ea" [ 1006.939424] env[63372]: _type = "HttpNfcLease" [ 1006.939424] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1006.950032] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1006.950032] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52623459-9c36-e714-df7d-df9c4b8282ea" [ 1006.950032] env[63372]: _type = "HttpNfcLease" [ 1006.950032] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1007.019700] env[63372]: DEBUG nova.network.neutron [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Successfully created port: 845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1007.076213] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb83938-726f-460d-8b17-911eb3242405 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.083625] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfe3cfb0-a3ad-4037-92d8-11df8aff4d10 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.122444] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce07fb2-fb65-4109-9a6d-96e861c1b193 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.130020] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.130365] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.130617] env[63372]: DEBUG nova.network.neutron [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1007.136118] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ad4f21-6f8a-481d-aea9-48f09bf165b0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.150944] env[63372]: DEBUG nova.compute.provider_tree [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.293686] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024641, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.564178} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.294158] env[63372]: INFO nova.virt.vmwareapi.ds_util [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 1724ce03-c8e2-415d-a380-59ac69fbfb57/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk. [ 1007.295013] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9effe00b-90e3-486d-93db-2bddd9bd9e33 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.322996] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 1724ce03-c8e2-415d-a380-59ac69fbfb57/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1007.323461] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e469a69d-c03d-44fa-8b2e-51126a7bc719 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.344083] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1007.344083] env[63372]: value = "task-1024643" [ 1007.344083] env[63372]: _type = "Task" [ 1007.344083] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.350856] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024643, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.449022] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1007.449022] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52623459-9c36-e714-df7d-df9c4b8282ea" [ 1007.449022] env[63372]: _type = "HttpNfcLease" [ 1007.449022] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1007.449429] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1007.449429] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52623459-9c36-e714-df7d-df9c4b8282ea" [ 1007.449429] env[63372]: _type = "HttpNfcLease" [ 1007.449429] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1007.450257] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54db06f8-976a-49b1-9bff-10d7adcb6301 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.458117] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524910d3-3651-60c9-5d0a-4207788274da/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1007.458314] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524910d3-3651-60c9-5d0a-4207788274da/disk-0.vmdk for reading. {{(pid=63372) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1007.534389] env[63372]: DEBUG nova.compute.manager [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Received event network-changed-301c1926-329c-454a-9755-e33bef4550d8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.534389] env[63372]: DEBUG nova.compute.manager [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Refreshing instance network info cache due to event network-changed-301c1926-329c-454a-9755-e33bef4550d8. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1007.534484] env[63372]: DEBUG oslo_concurrency.lockutils [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] Acquiring lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.535731] env[63372]: DEBUG oslo_concurrency.lockutils [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] Acquired lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.535731] env[63372]: DEBUG nova.network.neutron [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Refreshing network info cache for port 301c1926-329c-454a-9755-e33bef4550d8 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1007.556132] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a8f8358f-aa26-4486-a7e5-7f4761256f62 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.630530] env[63372]: DEBUG oslo_concurrency.lockutils [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "interface-b455b16b-3332-43bf-a91e-a350287ba5f3-1d17bf13-a03b-4654-bdfe-adc2d47e273a" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.630530] env[63372]: DEBUG oslo_concurrency.lockutils [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-b455b16b-3332-43bf-a91e-a350287ba5f3-1d17bf13-a03b-4654-bdfe-adc2d47e273a" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.630530] env[63372]: DEBUG nova.objects.instance [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'flavor' on Instance uuid b455b16b-3332-43bf-a91e-a350287ba5f3 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.654036] env[63372]: DEBUG nova.scheduler.client.report [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1007.678734] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "b7732621-7e58-40a3-b723-5c66df6f74a1" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.679058] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "b7732621-7e58-40a3-b723-5c66df6f74a1" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.679308] env[63372]: INFO nova.compute.manager [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Attaching volume 2d01e3c9-8f3a-4399-a8a8-213b5e212d9a to /dev/sdb [ 1007.683833] env[63372]: DEBUG nova.network.neutron [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1007.734728] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ef30f2-7fd5-4b24-87b6-552f16c440a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.749839] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635dac76-6d07-4da2-819c-b04d42832f84 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.753722] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.753905] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.754455] env[63372]: DEBUG nova.network.neutron [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1007.771870] env[63372]: DEBUG nova.virt.block_device [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating existing volume attachment record: 77a29910-dcdc-4a2e-981c-9c66bf8f18a4 {{(pid=63372) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1007.815970] env[63372]: DEBUG oslo_vmware.rw_handles [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52013295-109e-267e-17d8-e42f47fad683/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1007.817233] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b51dbe-caff-4605-8491-d2180f5abf3c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.824058] env[63372]: DEBUG oslo_vmware.rw_handles [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52013295-109e-267e-17d8-e42f47fad683/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1007.824271] env[63372]: ERROR oslo_vmware.rw_handles [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52013295-109e-267e-17d8-e42f47fad683/disk-0.vmdk due to incomplete transfer. [ 1007.824486] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-73b7db1b-f70f-48cf-8c94-741697ecd4af {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.831850] env[63372]: DEBUG oslo_vmware.rw_handles [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52013295-109e-267e-17d8-e42f47fad683/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1007.832099] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Uploaded image 243dbf27-41ba-48e9-ac46-97223c25bb0e to the Glance image server {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1007.834626] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1007.834931] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-958cc27a-d44d-46ce-9d7a-89fd122fae07 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.840749] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1007.840749] env[63372]: value = "task-1024644" [ 1007.840749] env[63372]: _type = "Task" [ 1007.840749] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.856544] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024644, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.860341] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024643, 'name': ReconfigVM_Task, 'duration_secs': 0.319407} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.860341] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 1724ce03-c8e2-415d-a380-59ac69fbfb57/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1007.860639] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea020cdc-e7ba-47b0-a215-89d6d17bda27 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.882981] env[63372]: DEBUG nova.compute.manager [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1007.891968] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20544fca-74f2-4ba6-abd6-959772ebba4e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.906156] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1007.906156] env[63372]: value = "task-1024645" [ 1007.906156] env[63372]: _type = "Task" [ 1007.906156] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.913604] env[63372]: DEBUG nova.network.neutron [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance_info_cache with network_info: [{"id": "7cd17ebc-3698-499b-8a5a-0507f100485c", "address": "fa:16:3e:59:5c:0e", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cd17ebc-36", "ovs_interfaceid": "7cd17ebc-3698-499b-8a5a-0507f100485c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.918275] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024645, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.920655] env[63372]: DEBUG nova.virt.hardware [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1007.920925] env[63372]: DEBUG nova.virt.hardware [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1007.921108] env[63372]: DEBUG nova.virt.hardware [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1007.921296] env[63372]: DEBUG nova.virt.hardware [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1007.921446] env[63372]: DEBUG nova.virt.hardware [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1007.921595] env[63372]: DEBUG nova.virt.hardware [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1007.921866] env[63372]: DEBUG nova.virt.hardware [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1007.922065] env[63372]: DEBUG nova.virt.hardware [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1007.922285] env[63372]: DEBUG nova.virt.hardware [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1007.922557] env[63372]: DEBUG nova.virt.hardware [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1007.922841] env[63372]: DEBUG nova.virt.hardware [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1007.923832] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920dcaa2-6ada-49d0-bd68-0087cd8e24bf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.932187] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23499bcb-1ac5-49a7-a08c-5978cbee85cc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.158139] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.314s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.161719] env[63372]: DEBUG nova.compute.manager [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1008.267133] env[63372]: DEBUG nova.objects.instance [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'pci_requests' on Instance uuid b455b16b-3332-43bf-a91e-a350287ba5f3 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.312671] env[63372]: DEBUG nova.network.neutron [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Updated VIF entry in instance network info cache for port 301c1926-329c-454a-9755-e33bef4550d8. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1008.313382] env[63372]: DEBUG nova.network.neutron [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Updating instance_info_cache with network_info: [{"id": "301c1926-329c-454a-9755-e33bef4550d8", "address": "fa:16:3e:bc:d9:91", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap301c1926-32", "ovs_interfaceid": "301c1926-329c-454a-9755-e33bef4550d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.355673] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024644, 'name': Destroy_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.418635] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024645, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.424402] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.424814] env[63372]: DEBUG nova.compute.manager [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Instance network_info: |[{"id": "7cd17ebc-3698-499b-8a5a-0507f100485c", "address": "fa:16:3e:59:5c:0e", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cd17ebc-36", "ovs_interfaceid": "7cd17ebc-3698-499b-8a5a-0507f100485c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1008.425195] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:5c:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d5d0e0d-cdec-474a-a891-a9ceff15a8b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7cd17ebc-3698-499b-8a5a-0507f100485c', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1008.434314] env[63372]: DEBUG oslo.service.loopingcall [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.434585] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1008.434986] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d1970b39-d919-46f2-8835-0a3a2fb9530b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.457570] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1008.457570] env[63372]: value = "task-1024649" [ 1008.457570] env[63372]: _type = "Task" [ 1008.457570] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.466355] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024649, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.491224] env[63372]: DEBUG nova.network.neutron [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating instance_info_cache with network_info: [{"id": "298d217f-c17a-4ee2-acae-baee355112dd", "address": "fa:16:3e:66:af:86", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298d217f-c1", "ovs_interfaceid": "298d217f-c17a-4ee2-acae-baee355112dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.669952] env[63372]: DEBUG nova.compute.utils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1008.677217] env[63372]: DEBUG nova.compute.manager [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1008.677217] env[63372]: DEBUG nova.network.neutron [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1008.720927] env[63372]: DEBUG nova.policy [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb233d4b6fe54fefae3046d62d167e34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d0c0f315749429dbae22a72ae82b1b2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1008.767227] env[63372]: DEBUG nova.objects.base [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1008.767888] env[63372]: DEBUG nova.network.neutron [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1008.816194] env[63372]: DEBUG oslo_concurrency.lockutils [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] Releasing lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.816492] env[63372]: DEBUG nova.compute.manager [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Received event network-changed-301c1926-329c-454a-9755-e33bef4550d8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.816663] env[63372]: DEBUG nova.compute.manager [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Refreshing instance network info cache due to event network-changed-301c1926-329c-454a-9755-e33bef4550d8. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1008.816883] env[63372]: DEBUG oslo_concurrency.lockutils [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] Acquiring lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.817045] env[63372]: DEBUG oslo_concurrency.lockutils [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] Acquired lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.817214] env[63372]: DEBUG nova.network.neutron [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Refreshing network info cache for port 301c1926-329c-454a-9755-e33bef4550d8 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1008.854677] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024644, 'name': Destroy_Task, 'duration_secs': 0.836098} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.855074] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Destroyed the VM [ 1008.855330] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1008.855864] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d30e97d4-5818-49e8-a26b-026845a58e84 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.862710] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1008.862710] env[63372]: value = "task-1024650" [ 1008.862710] env[63372]: _type = "Task" [ 1008.862710] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.870464] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024650, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.919275] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024645, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.967458] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024649, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.994230] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.002256] env[63372]: DEBUG nova.network.neutron [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Successfully created port: f96d5780-3a85-4214-8d8e-2b02ca1c4c2e {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1009.046470] env[63372]: DEBUG nova.policy [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7ffe6776b664c00b9086046f9171ec6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46d2cb188a004dc385f2aa15458be929', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1009.078219] env[63372]: DEBUG nova.network.neutron [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Successfully updated port: 845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1009.176811] env[63372]: DEBUG nova.compute.manager [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1009.297535] env[63372]: DEBUG nova.compute.manager [req-7e4a1f75-c991-4fff-8d7c-12c3ff9b74a8 req-6db765f2-4088-4e3c-b7a2-a883ae7ae207 service nova] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Received event network-changed-7cd17ebc-3698-499b-8a5a-0507f100485c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.297789] env[63372]: DEBUG nova.compute.manager [req-7e4a1f75-c991-4fff-8d7c-12c3ff9b74a8 req-6db765f2-4088-4e3c-b7a2-a883ae7ae207 service nova] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Refreshing instance network info cache due to event network-changed-7cd17ebc-3698-499b-8a5a-0507f100485c. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1009.298016] env[63372]: DEBUG oslo_concurrency.lockutils [req-7e4a1f75-c991-4fff-8d7c-12c3ff9b74a8 req-6db765f2-4088-4e3c-b7a2-a883ae7ae207 service nova] Acquiring lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.298172] env[63372]: DEBUG oslo_concurrency.lockutils [req-7e4a1f75-c991-4fff-8d7c-12c3ff9b74a8 req-6db765f2-4088-4e3c-b7a2-a883ae7ae207 service nova] Acquired lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.298338] env[63372]: DEBUG nova.network.neutron [req-7e4a1f75-c991-4fff-8d7c-12c3ff9b74a8 req-6db765f2-4088-4e3c-b7a2-a883ae7ae207 service nova] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Refreshing network info cache for port 7cd17ebc-3698-499b-8a5a-0507f100485c {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1009.374553] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024650, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.418972] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024645, 'name': ReconfigVM_Task, 'duration_secs': 1.26471} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.419256] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1009.419637] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f0c9c03-8ffb-4196-8095-9b8806dce1ef {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.426545] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1009.426545] env[63372]: value = "task-1024651" [ 1009.426545] env[63372]: _type = "Task" [ 1009.426545] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.435576] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024651, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.467312] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024649, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.526087] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8c83fc-caf3-4d8f-b3f5-39bf5e1030ef {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.548636] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-275b062d-87e9-45e4-8639-1873c00f102c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.556238] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating instance 'cbb471c5-8d86-4d8f-8935-f6a77e5144be' progress to 83 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1009.586022] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.586022] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.586022] env[63372]: DEBUG nova.network.neutron [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1009.729396] env[63372]: DEBUG nova.network.neutron [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Updated VIF entry in instance network info cache for port 301c1926-329c-454a-9755-e33bef4550d8. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1009.729835] env[63372]: DEBUG nova.network.neutron [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Updating instance_info_cache with network_info: [{"id": "301c1926-329c-454a-9755-e33bef4550d8", "address": "fa:16:3e:bc:d9:91", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap301c1926-32", "ovs_interfaceid": "301c1926-329c-454a-9755-e33bef4550d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.875477] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024650, 'name': RemoveSnapshot_Task, 'duration_secs': 0.668327} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.875697] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1009.875963] env[63372]: DEBUG nova.compute.manager [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1009.876805] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cee73c7-cc16-4b63-bb01-0b77d26f774e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.938578] env[63372]: DEBUG oslo_vmware.api [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024651, 'name': PowerOnVM_Task, 'duration_secs': 0.481702} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.939089] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1009.942243] env[63372]: DEBUG nova.compute.manager [None req-78f123b5-3ed0-40f5-9b31-17ab8e8da33c tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1009.943150] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f426534-945c-4b5b-bc9c-55b784f094ea {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.971782] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024649, 'name': CreateVM_Task, 'duration_secs': 1.012215} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.971782] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1009.971782] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.971782] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.971782] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1009.971782] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1218bea1-8098-42a0-a21d-96c196b08fb1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.977177] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1009.977177] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52de080c-7379-9ae3-8d5a-31a85eb17c0a" [ 1009.977177] env[63372]: _type = "Task" [ 1009.977177] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.985626] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52de080c-7379-9ae3-8d5a-31a85eb17c0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.062893] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1010.065491] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d86deec-db09-496a-953f-5b2e199af1b4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.073243] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1010.073243] env[63372]: value = "task-1024652" [ 1010.073243] env[63372]: _type = "Task" [ 1010.073243] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.081825] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024652, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.114406] env[63372]: DEBUG nova.network.neutron [req-7e4a1f75-c991-4fff-8d7c-12c3ff9b74a8 req-6db765f2-4088-4e3c-b7a2-a883ae7ae207 service nova] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updated VIF entry in instance network info cache for port 7cd17ebc-3698-499b-8a5a-0507f100485c. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1010.114860] env[63372]: DEBUG nova.network.neutron [req-7e4a1f75-c991-4fff-8d7c-12c3ff9b74a8 req-6db765f2-4088-4e3c-b7a2-a883ae7ae207 service nova] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance_info_cache with network_info: [{"id": "7cd17ebc-3698-499b-8a5a-0507f100485c", "address": "fa:16:3e:59:5c:0e", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cd17ebc-36", "ovs_interfaceid": "7cd17ebc-3698-499b-8a5a-0507f100485c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.118751] env[63372]: DEBUG nova.network.neutron [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1010.190928] env[63372]: DEBUG nova.compute.manager [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1010.216478] env[63372]: DEBUG nova.virt.hardware [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1010.216751] env[63372]: DEBUG nova.virt.hardware [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1010.216917] env[63372]: DEBUG nova.virt.hardware [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1010.217122] env[63372]: DEBUG nova.virt.hardware [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1010.217274] env[63372]: DEBUG nova.virt.hardware [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1010.217423] env[63372]: DEBUG nova.virt.hardware [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1010.217637] env[63372]: DEBUG nova.virt.hardware [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1010.217785] env[63372]: DEBUG nova.virt.hardware [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1010.217955] env[63372]: DEBUG nova.virt.hardware [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1010.218135] env[63372]: DEBUG nova.virt.hardware [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1010.218320] env[63372]: DEBUG nova.virt.hardware [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1010.219251] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c83535-443e-4b10-94f6-77fa644ee8ac {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.229644] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b4d001a-591f-4220-a375-a46f6fa75726 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.233768] env[63372]: DEBUG oslo_concurrency.lockutils [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] Releasing lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.234035] env[63372]: DEBUG nova.compute.manager [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Received event network-changed-18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.234235] env[63372]: DEBUG nova.compute.manager [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Refreshing instance network info cache due to event network-changed-18c7850a-d97e-4564-92e5-38d1204369e4. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1010.234445] env[63372]: DEBUG oslo_concurrency.lockutils [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] Acquiring lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.234589] env[63372]: DEBUG oslo_concurrency.lockutils [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] Acquired lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.234752] env[63372]: DEBUG nova.network.neutron [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Refreshing network info cache for port 18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1010.256499] env[63372]: DEBUG nova.network.neutron [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Updating instance_info_cache with network_info: [{"id": "845b30fa-3e7e-4324-9479-aed7aade3f91", "address": "fa:16:3e:54:d5:15", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap845b30fa-3e", "ovs_interfaceid": "845b30fa-3e7e-4324-9479-aed7aade3f91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.390901] env[63372]: INFO nova.compute.manager [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Shelve offloading [ 1010.392376] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1010.392701] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed2534e9-09b6-4dea-a9b4-ffbf541a8a42 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.399776] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1010.399776] env[63372]: value = "task-1024654" [ 1010.399776] env[63372]: _type = "Task" [ 1010.399776] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.409608] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024654, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.435935] env[63372]: DEBUG nova.compute.manager [req-5a94cf0a-904e-425e-b0b3-68481a3bb27f req-a7a1c5bf-b470-4885-8741-acd006f1f29a service nova] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Received event network-vif-plugged-f96d5780-3a85-4214-8d8e-2b02ca1c4c2e {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.436187] env[63372]: DEBUG oslo_concurrency.lockutils [req-5a94cf0a-904e-425e-b0b3-68481a3bb27f req-a7a1c5bf-b470-4885-8741-acd006f1f29a service nova] Acquiring lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.436392] env[63372]: DEBUG oslo_concurrency.lockutils [req-5a94cf0a-904e-425e-b0b3-68481a3bb27f req-a7a1c5bf-b470-4885-8741-acd006f1f29a service nova] Lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.436573] env[63372]: DEBUG oslo_concurrency.lockutils [req-5a94cf0a-904e-425e-b0b3-68481a3bb27f req-a7a1c5bf-b470-4885-8741-acd006f1f29a service nova] Lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.436751] env[63372]: DEBUG nova.compute.manager [req-5a94cf0a-904e-425e-b0b3-68481a3bb27f req-a7a1c5bf-b470-4885-8741-acd006f1f29a service nova] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] No waiting events found dispatching network-vif-plugged-f96d5780-3a85-4214-8d8e-2b02ca1c4c2e {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1010.436954] env[63372]: WARNING nova.compute.manager [req-5a94cf0a-904e-425e-b0b3-68481a3bb27f req-a7a1c5bf-b470-4885-8741-acd006f1f29a service nova] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Received unexpected event network-vif-plugged-f96d5780-3a85-4214-8d8e-2b02ca1c4c2e for instance with vm_state building and task_state spawning. [ 1010.488807] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52de080c-7379-9ae3-8d5a-31a85eb17c0a, 'name': SearchDatastore_Task, 'duration_secs': 0.039488} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.489368] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.489701] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.490033] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.490277] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.490530] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.491218] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4eb70151-4182-4597-ba79-7eed993f5896 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.501527] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.501879] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1010.503085] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d64d576-35a0-4dd1-9735-62c3fef82ac8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.509092] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1010.509092] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52a5e7b2-4637-75ff-c090-1c286296ad7f" [ 1010.509092] env[63372]: _type = "Task" [ 1010.509092] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.517343] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a5e7b2-4637-75ff-c090-1c286296ad7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.543570] env[63372]: DEBUG nova.network.neutron [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Successfully updated port: f96d5780-3a85-4214-8d8e-2b02ca1c4c2e {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1010.582948] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024652, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.618075] env[63372]: DEBUG oslo_concurrency.lockutils [req-7e4a1f75-c991-4fff-8d7c-12c3ff9b74a8 req-6db765f2-4088-4e3c-b7a2-a883ae7ae207 service nova] Releasing lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.618075] env[63372]: DEBUG nova.compute.manager [req-7e4a1f75-c991-4fff-8d7c-12c3ff9b74a8 req-6db765f2-4088-4e3c-b7a2-a883ae7ae207 service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Received event network-vif-plugged-845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.618075] env[63372]: DEBUG oslo_concurrency.lockutils [req-7e4a1f75-c991-4fff-8d7c-12c3ff9b74a8 req-6db765f2-4088-4e3c-b7a2-a883ae7ae207 service nova] Acquiring lock "28eb1675-b1b8-46a6-873d-5c858b716575-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.618317] env[63372]: DEBUG oslo_concurrency.lockutils [req-7e4a1f75-c991-4fff-8d7c-12c3ff9b74a8 req-6db765f2-4088-4e3c-b7a2-a883ae7ae207 service nova] Lock "28eb1675-b1b8-46a6-873d-5c858b716575-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.618395] env[63372]: DEBUG oslo_concurrency.lockutils [req-7e4a1f75-c991-4fff-8d7c-12c3ff9b74a8 req-6db765f2-4088-4e3c-b7a2-a883ae7ae207 service nova] Lock "28eb1675-b1b8-46a6-873d-5c858b716575-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.618563] env[63372]: DEBUG nova.compute.manager [req-7e4a1f75-c991-4fff-8d7c-12c3ff9b74a8 req-6db765f2-4088-4e3c-b7a2-a883ae7ae207 service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] No waiting events found dispatching network-vif-plugged-845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1010.618879] env[63372]: WARNING nova.compute.manager [req-7e4a1f75-c991-4fff-8d7c-12c3ff9b74a8 req-6db765f2-4088-4e3c-b7a2-a883ae7ae207 service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Received unexpected event network-vif-plugged-845b30fa-3e7e-4324-9479-aed7aade3f91 for instance with vm_state building and task_state spawning. [ 1010.758642] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.759070] env[63372]: DEBUG nova.compute.manager [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Instance network_info: |[{"id": "845b30fa-3e7e-4324-9479-aed7aade3f91", "address": "fa:16:3e:54:d5:15", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap845b30fa-3e", "ovs_interfaceid": "845b30fa-3e7e-4324-9479-aed7aade3f91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1010.759458] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:d5:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '845b30fa-3e7e-4324-9479-aed7aade3f91', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1010.768564] env[63372]: DEBUG oslo.service.loopingcall [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.769213] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1010.769481] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6697f9ff-f1b8-412b-9332-7213a105dcd3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.792264] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1010.792264] env[63372]: value = "task-1024655" [ 1010.792264] env[63372]: _type = "Task" [ 1010.792264] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.800345] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024655, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.912391] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] VM already powered off {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1010.912391] env[63372]: DEBUG nova.compute.manager [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1010.913763] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080ee08b-e5e4-4d14-9f98-894c8d6a92f1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.922765] env[63372]: DEBUG oslo_concurrency.lockutils [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "refresh_cache-9be70e66-2d81-4c7b-8dcc-e69815faef1a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.922945] env[63372]: DEBUG oslo_concurrency.lockutils [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "refresh_cache-9be70e66-2d81-4c7b-8dcc-e69815faef1a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.923161] env[63372]: DEBUG nova.network.neutron [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1010.956692] env[63372]: DEBUG nova.network.neutron [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Successfully updated port: 1d17bf13-a03b-4654-bdfe-adc2d47e273a {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1011.005053] env[63372]: DEBUG nova.network.neutron [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updated VIF entry in instance network info cache for port 18c7850a-d97e-4564-92e5-38d1204369e4. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1011.005277] env[63372]: DEBUG nova.network.neutron [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updating instance_info_cache with network_info: [{"id": "18c7850a-d97e-4564-92e5-38d1204369e4", "address": "fa:16:3e:14:28:4a", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c7850a-d9", "ovs_interfaceid": "18c7850a-d97e-4564-92e5-38d1204369e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.020411] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a5e7b2-4637-75ff-c090-1c286296ad7f, 'name': SearchDatastore_Task, 'duration_secs': 0.01552} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.021257] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4f71e27-ab7b-4380-a736-cddd9b765782 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.026635] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1011.026635] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b9c23d-c50a-cc38-b6c1-24a9e236b134" [ 1011.026635] env[63372]: _type = "Task" [ 1011.026635] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.036534] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b9c23d-c50a-cc38-b6c1-24a9e236b134, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.049608] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "refresh_cache-800cb48b-e75e-45e7-94dc-6ad147e6c6c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.049685] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "refresh_cache-800cb48b-e75e-45e7-94dc-6ad147e6c6c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.049824] env[63372]: DEBUG nova.network.neutron [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1011.083827] env[63372]: DEBUG oslo_vmware.api [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024652, 'name': PowerOnVM_Task, 'duration_secs': 0.590178} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.084166] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1011.084365] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7d1bed93-20b6-4b51-9b2a-db092207c376 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating instance 'cbb471c5-8d86-4d8f-8935-f6a77e5144be' progress to 100 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1011.305074] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024655, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.460228] env[63372]: DEBUG oslo_concurrency.lockutils [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.463765] env[63372]: DEBUG nova.compute.manager [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Received event network-changed-845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.464067] env[63372]: DEBUG nova.compute.manager [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Refreshing instance network info cache due to event network-changed-845b30fa-3e7e-4324-9479-aed7aade3f91. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1011.464195] env[63372]: DEBUG oslo_concurrency.lockutils [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] Acquiring lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.464321] env[63372]: DEBUG oslo_concurrency.lockutils [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] Acquired lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.464518] env[63372]: DEBUG nova.network.neutron [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Refreshing network info cache for port 845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1011.509508] env[63372]: DEBUG oslo_concurrency.lockutils [req-9024ca28-f58b-4b9b-9ce5-787cc095003e req-4903a37d-cf54-46d7-80e9-59e1c00efe54 service nova] Releasing lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.510020] env[63372]: DEBUG oslo_concurrency.lockutils [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.510236] env[63372]: DEBUG nova.network.neutron [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1011.537445] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b9c23d-c50a-cc38-b6c1-24a9e236b134, 'name': SearchDatastore_Task, 'duration_secs': 0.018513} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.537612] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.538042] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 4a2e53e0-b68c-42be-afff-1eead017679d/4a2e53e0-b68c-42be-afff-1eead017679d.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1011.538178] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb4946f4-522d-41f9-8cfa-a8babec34854 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.544557] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1011.544557] env[63372]: value = "task-1024656" [ 1011.544557] env[63372]: _type = "Task" [ 1011.544557] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.556491] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024656, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.593565] env[63372]: DEBUG nova.network.neutron [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1011.660320] env[63372]: DEBUG nova.network.neutron [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Updating instance_info_cache with network_info: [{"id": "9c53b112-6225-4e7e-95d5-cce3c462a615", "address": "fa:16:3e:70:bf:65", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c53b112-62", "ovs_interfaceid": "9c53b112-6225-4e7e-95d5-cce3c462a615", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.773822] env[63372]: DEBUG nova.network.neutron [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Updating instance_info_cache with network_info: [{"id": "f96d5780-3a85-4214-8d8e-2b02ca1c4c2e", "address": "fa:16:3e:db:19:3d", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf96d5780-3a", "ovs_interfaceid": "f96d5780-3a85-4214-8d8e-2b02ca1c4c2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.803675] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024655, 'name': CreateVM_Task, 'duration_secs': 0.519857} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.803848] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1011.804529] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.804698] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.805025] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1011.805280] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c5f891f-0be7-4ad4-b4ee-f244058863d2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.810121] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1011.810121] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b2cbcf-2616-09ca-047f-3b3df85c842f" [ 1011.810121] env[63372]: _type = "Task" [ 1011.810121] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.818154] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b2cbcf-2616-09ca-047f-3b3df85c842f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.045154] env[63372]: WARNING nova.network.neutron [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] e1f7ca1b-541e-468f-a439-a6841fc781ba already exists in list: networks containing: ['e1f7ca1b-541e-468f-a439-a6841fc781ba']. ignoring it [ 1012.057015] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024656, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.165180] env[63372]: DEBUG oslo_concurrency.lockutils [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "refresh_cache-9be70e66-2d81-4c7b-8dcc-e69815faef1a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.243897] env[63372]: DEBUG nova.network.neutron [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Updated VIF entry in instance network info cache for port 845b30fa-3e7e-4324-9479-aed7aade3f91. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1012.244315] env[63372]: DEBUG nova.network.neutron [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Updating instance_info_cache with network_info: [{"id": "845b30fa-3e7e-4324-9479-aed7aade3f91", "address": "fa:16:3e:54:d5:15", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap845b30fa-3e", "ovs_interfaceid": "845b30fa-3e7e-4324-9479-aed7aade3f91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.276837] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "refresh_cache-800cb48b-e75e-45e7-94dc-6ad147e6c6c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.277201] env[63372]: DEBUG nova.compute.manager [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Instance network_info: |[{"id": "f96d5780-3a85-4214-8d8e-2b02ca1c4c2e", "address": "fa:16:3e:db:19:3d", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf96d5780-3a", "ovs_interfaceid": "f96d5780-3a85-4214-8d8e-2b02ca1c4c2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1012.280331] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:19:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60badc2d-69d2-467d-a92e-98511f5cb0b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f96d5780-3a85-4214-8d8e-2b02ca1c4c2e', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1012.289478] env[63372]: DEBUG oslo.service.loopingcall [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.292341] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1012.292974] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ad19f22-40b3-4d2f-9977-adc52b04624f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.316178] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1012.316178] env[63372]: value = "task-1024657" [ 1012.316178] env[63372]: _type = "Task" [ 1012.316178] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.326898] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b2cbcf-2616-09ca-047f-3b3df85c842f, 'name': SearchDatastore_Task, 'duration_secs': 0.022285} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.326898] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.326898] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1012.326898] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.326898] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.326898] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1012.331435] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d4e3881-9efb-4365-b6b9-66ea9f89e80b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.334347] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024657, 'name': CreateVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.348032] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1012.348164] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1012.356642] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c76b982-e87f-4297-9c95-3cb000a5a91b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.356642] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1012.356642] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b47847-54d3-e11c-56a6-ce03300a31ba" [ 1012.356642] env[63372]: _type = "Task" [ 1012.356642] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.367991] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b47847-54d3-e11c-56a6-ce03300a31ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.549040] env[63372]: DEBUG nova.network.neutron [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updating instance_info_cache with network_info: [{"id": "18c7850a-d97e-4564-92e5-38d1204369e4", "address": "fa:16:3e:14:28:4a", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c7850a-d9", "ovs_interfaceid": "18c7850a-d97e-4564-92e5-38d1204369e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1d17bf13-a03b-4654-bdfe-adc2d47e273a", "address": "fa:16:3e:09:36:65", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d17bf13-a0", "ovs_interfaceid": "1d17bf13-a03b-4654-bdfe-adc2d47e273a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.559724] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024656, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.817454} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.560016] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 4a2e53e0-b68c-42be-afff-1eead017679d/4a2e53e0-b68c-42be-afff-1eead017679d.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1012.560250] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1012.560509] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2ef7a2d2-fe74-4add-90b4-ee38e2f50a9a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.567904] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1012.567904] env[63372]: value = "task-1024658" [ 1012.567904] env[63372]: _type = "Task" [ 1012.567904] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.578412] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024658, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.630119] env[63372]: DEBUG nova.compute.manager [req-c28d56ad-ad08-4e95-b16f-a3e08bc40018 req-34952a6f-6c60-4fcf-a63e-25fa1abde198 service nova] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Received event network-changed-f96d5780-3a85-4214-8d8e-2b02ca1c4c2e {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.630356] env[63372]: DEBUG nova.compute.manager [req-c28d56ad-ad08-4e95-b16f-a3e08bc40018 req-34952a6f-6c60-4fcf-a63e-25fa1abde198 service nova] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Refreshing instance network info cache due to event network-changed-f96d5780-3a85-4214-8d8e-2b02ca1c4c2e. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1012.630597] env[63372]: DEBUG oslo_concurrency.lockutils [req-c28d56ad-ad08-4e95-b16f-a3e08bc40018 req-34952a6f-6c60-4fcf-a63e-25fa1abde198 service nova] Acquiring lock "refresh_cache-800cb48b-e75e-45e7-94dc-6ad147e6c6c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.630750] env[63372]: DEBUG oslo_concurrency.lockutils [req-c28d56ad-ad08-4e95-b16f-a3e08bc40018 req-34952a6f-6c60-4fcf-a63e-25fa1abde198 service nova] Acquired lock "refresh_cache-800cb48b-e75e-45e7-94dc-6ad147e6c6c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.630983] env[63372]: DEBUG nova.network.neutron [req-c28d56ad-ad08-4e95-b16f-a3e08bc40018 req-34952a6f-6c60-4fcf-a63e-25fa1abde198 service nova] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Refreshing network info cache for port f96d5780-3a85-4214-8d8e-2b02ca1c4c2e {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1012.645330] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1012.646477] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6cac4aa-4616-4a38-bd8b-98e11b281941 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.654994] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1012.655319] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ecc38505-2341-42fa-a5c2-220b4a20f37f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.748356] env[63372]: DEBUG oslo_concurrency.lockutils [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] Releasing lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.748798] env[63372]: DEBUG nova.compute.manager [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Received event network-vif-plugged-1d17bf13-a03b-4654-bdfe-adc2d47e273a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.749034] env[63372]: DEBUG oslo_concurrency.lockutils [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] Acquiring lock "b455b16b-3332-43bf-a91e-a350287ba5f3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.749367] env[63372]: DEBUG oslo_concurrency.lockutils [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] Lock "b455b16b-3332-43bf-a91e-a350287ba5f3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.749563] env[63372]: DEBUG oslo_concurrency.lockutils [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] Lock "b455b16b-3332-43bf-a91e-a350287ba5f3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.749747] env[63372]: DEBUG nova.compute.manager [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] No waiting events found dispatching network-vif-plugged-1d17bf13-a03b-4654-bdfe-adc2d47e273a {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1012.750020] env[63372]: WARNING nova.compute.manager [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Received unexpected event network-vif-plugged-1d17bf13-a03b-4654-bdfe-adc2d47e273a for instance with vm_state active and task_state None. [ 1012.750258] env[63372]: DEBUG nova.compute.manager [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Received event network-changed-1d17bf13-a03b-4654-bdfe-adc2d47e273a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.750436] env[63372]: DEBUG nova.compute.manager [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Refreshing instance network info cache due to event network-changed-1d17bf13-a03b-4654-bdfe-adc2d47e273a. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1012.750612] env[63372]: DEBUG oslo_concurrency.lockutils [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] Acquiring lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.823525] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1012.823760] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1012.823947] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleting the datastore file [datastore2] 9be70e66-2d81-4c7b-8dcc-e69815faef1a {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1012.824991] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Volume attach. Driver type: vmdk {{(pid=63372) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1012.825250] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227445', 'volume_id': '2d01e3c9-8f3a-4399-a8a8-213b5e212d9a', 'name': 'volume-2d01e3c9-8f3a-4399-a8a8-213b5e212d9a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b7732621-7e58-40a3-b723-5c66df6f74a1', 'attached_at': '', 'detached_at': '', 'volume_id': '2d01e3c9-8f3a-4399-a8a8-213b5e212d9a', 'serial': '2d01e3c9-8f3a-4399-a8a8-213b5e212d9a'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1012.825892] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9183924d-3111-4f67-b138-f8df9a7fc123 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.828313] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd65697-9891-4793-a613-0dc191d258a8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.834107] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024657, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.850804] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d948113a-aa25-43d5-a934-8a47621dac0d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.853600] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1012.853600] env[63372]: value = "task-1024660" [ 1012.853600] env[63372]: _type = "Task" [ 1012.853600] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.878048] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] volume-2d01e3c9-8f3a-4399-a8a8-213b5e212d9a/volume-2d01e3c9-8f3a-4399-a8a8-213b5e212d9a.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1012.881922] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed74e0d9-96ee-46a7-a253-a17bef72e1a0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.899190] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024660, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.905549] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b47847-54d3-e11c-56a6-ce03300a31ba, 'name': SearchDatastore_Task, 'duration_secs': 0.012157} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.907499] env[63372]: DEBUG oslo_vmware.api [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1012.907499] env[63372]: value = "task-1024661" [ 1012.907499] env[63372]: _type = "Task" [ 1012.907499] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.907704] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f75b24f-ccca-4b70-b18f-e00818e6054b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.915988] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1012.915988] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52dd2489-f540-9780-4d4a-37cdc24a03dd" [ 1012.915988] env[63372]: _type = "Task" [ 1012.915988] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.919149] env[63372]: DEBUG oslo_vmware.api [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024661, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.926456] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52dd2489-f540-9780-4d4a-37cdc24a03dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.055332] env[63372]: DEBUG oslo_concurrency.lockutils [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.056054] env[63372]: DEBUG oslo_concurrency.lockutils [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.056215] env[63372]: DEBUG oslo_concurrency.lockutils [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.056524] env[63372]: DEBUG oslo_concurrency.lockutils [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] Acquired lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.056753] env[63372]: DEBUG nova.network.neutron [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Refreshing network info cache for port 1d17bf13-a03b-4654-bdfe-adc2d47e273a {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1013.059037] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5bea8ce-a8ab-4420-aedf-638ee1a1f483 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.075932] env[63372]: DEBUG nova.virt.hardware [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.076162] env[63372]: DEBUG nova.virt.hardware [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.076320] env[63372]: DEBUG nova.virt.hardware [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.076504] env[63372]: DEBUG nova.virt.hardware [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.076649] env[63372]: DEBUG nova.virt.hardware [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.076797] env[63372]: DEBUG nova.virt.hardware [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.076999] env[63372]: DEBUG nova.virt.hardware [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.077173] env[63372]: DEBUG nova.virt.hardware [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.077333] env[63372]: DEBUG nova.virt.hardware [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.077494] env[63372]: DEBUG nova.virt.hardware [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.077662] env[63372]: DEBUG nova.virt.hardware [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.083931] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Reconfiguring VM to attach interface {{(pid=63372) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1013.087648] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a5b461d-974a-467a-9ee7-d54f9bda509f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.105323] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024658, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066288} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.106582] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1013.106931] env[63372]: DEBUG oslo_vmware.api [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 1013.106931] env[63372]: value = "task-1024662" [ 1013.106931] env[63372]: _type = "Task" [ 1013.106931] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.107622] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd9e057-1ea2-4153-adb6-444ae64f09ba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.119079] env[63372]: DEBUG oslo_vmware.api [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024662, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.136250] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 4a2e53e0-b68c-42be-afff-1eead017679d/4a2e53e0-b68c-42be-afff-1eead017679d.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.138874] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8454020-427d-4b52-bd22-c310be895265 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.158507] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1013.158507] env[63372]: value = "task-1024663" [ 1013.158507] env[63372]: _type = "Task" [ 1013.158507] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.169283] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024663, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.295838] env[63372]: INFO nova.compute.manager [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Rescuing [ 1013.296100] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.296266] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.296437] env[63372]: DEBUG nova.network.neutron [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1013.328209] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024657, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.363663] env[63372]: DEBUG oslo_vmware.api [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024660, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.37246} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.364418] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1013.364418] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1013.364539] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1013.369197] env[63372]: DEBUG nova.network.neutron [req-c28d56ad-ad08-4e95-b16f-a3e08bc40018 req-34952a6f-6c60-4fcf-a63e-25fa1abde198 service nova] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Updated VIF entry in instance network info cache for port f96d5780-3a85-4214-8d8e-2b02ca1c4c2e. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1013.369613] env[63372]: DEBUG nova.network.neutron [req-c28d56ad-ad08-4e95-b16f-a3e08bc40018 req-34952a6f-6c60-4fcf-a63e-25fa1abde198 service nova] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Updating instance_info_cache with network_info: [{"id": "f96d5780-3a85-4214-8d8e-2b02ca1c4c2e", "address": "fa:16:3e:db:19:3d", "network": {"id": "e31a802a-2998-4f2f-a812-af03fb5ec8c1", "bridge": "br-int", "label": "tempest-ServersTestJSON-286962482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d0c0f315749429dbae22a72ae82b1b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf96d5780-3a", "ovs_interfaceid": "f96d5780-3a85-4214-8d8e-2b02ca1c4c2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.404486] env[63372]: INFO nova.scheduler.client.report [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleted allocations for instance 9be70e66-2d81-4c7b-8dcc-e69815faef1a [ 1013.421459] env[63372]: DEBUG oslo_vmware.api [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024661, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.430607] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52dd2489-f540-9780-4d4a-37cdc24a03dd, 'name': SearchDatastore_Task, 'duration_secs': 0.019892} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.430961] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.431577] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 28eb1675-b1b8-46a6-873d-5c858b716575/28eb1675-b1b8-46a6-873d-5c858b716575.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1013.431711] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ab137a04-2134-447d-a176-15e46db13767 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.439457] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1013.439457] env[63372]: value = "task-1024664" [ 1013.439457] env[63372]: _type = "Task" [ 1013.439457] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.451410] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024664, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.619568] env[63372]: DEBUG oslo_vmware.api [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024662, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.668014] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024663, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.774202] env[63372]: DEBUG nova.network.neutron [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updated VIF entry in instance network info cache for port 1d17bf13-a03b-4654-bdfe-adc2d47e273a. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1013.774644] env[63372]: DEBUG nova.network.neutron [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updating instance_info_cache with network_info: [{"id": "18c7850a-d97e-4564-92e5-38d1204369e4", "address": "fa:16:3e:14:28:4a", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c7850a-d9", "ovs_interfaceid": "18c7850a-d97e-4564-92e5-38d1204369e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1d17bf13-a03b-4654-bdfe-adc2d47e273a", "address": "fa:16:3e:09:36:65", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d17bf13-a0", "ovs_interfaceid": "1d17bf13-a03b-4654-bdfe-adc2d47e273a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.829050] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024657, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.852960] env[63372]: DEBUG oslo_concurrency.lockutils [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.853241] env[63372]: DEBUG oslo_concurrency.lockutils [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.853431] env[63372]: DEBUG nova.compute.manager [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Going to confirm migration 4 {{(pid=63372) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1013.872614] env[63372]: DEBUG oslo_concurrency.lockutils [req-c28d56ad-ad08-4e95-b16f-a3e08bc40018 req-34952a6f-6c60-4fcf-a63e-25fa1abde198 service nova] Releasing lock "refresh_cache-800cb48b-e75e-45e7-94dc-6ad147e6c6c5" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.909943] env[63372]: DEBUG oslo_concurrency.lockutils [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.910243] env[63372]: DEBUG oslo_concurrency.lockutils [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.910480] env[63372]: DEBUG nova.objects.instance [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lazy-loading 'resources' on Instance uuid 9be70e66-2d81-4c7b-8dcc-e69815faef1a {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.932108] env[63372]: DEBUG oslo_vmware.api [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024661, 'name': ReconfigVM_Task, 'duration_secs': 0.726346} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.932453] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Reconfigured VM instance instance-0000005a to attach disk [datastore1] volume-2d01e3c9-8f3a-4399-a8a8-213b5e212d9a/volume-2d01e3c9-8f3a-4399-a8a8-213b5e212d9a.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.940024] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2fdf386e-3c5a-486b-99cb-7137eea58c70 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.964442] env[63372]: DEBUG oslo_vmware.api [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1013.964442] env[63372]: value = "task-1024665" [ 1013.964442] env[63372]: _type = "Task" [ 1013.964442] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.968390] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024664, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.979348] env[63372]: DEBUG oslo_vmware.api [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024665, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.087336] env[63372]: DEBUG nova.network.neutron [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Updating instance_info_cache with network_info: [{"id": "b54d2aaa-a37b-4db7-b735-f9d2a6a9534b", "address": "fa:16:3e:03:b6:cc", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb54d2aaa-a3", "ovs_interfaceid": "b54d2aaa-a37b-4db7-b735-f9d2a6a9534b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.119962] env[63372]: DEBUG oslo_vmware.api [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024662, 'name': ReconfigVM_Task, 'duration_secs': 0.859928} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.120495] env[63372]: DEBUG oslo_concurrency.lockutils [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.120713] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Reconfigured VM to attach interface {{(pid=63372) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1014.169040] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024663, 'name': ReconfigVM_Task, 'duration_secs': 0.5216} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.169040] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 4a2e53e0-b68c-42be-afff-1eead017679d/4a2e53e0-b68c-42be-afff-1eead017679d.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.169794] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f578020-350c-4f3b-81de-4988098eb45e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.176390] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1014.176390] env[63372]: value = "task-1024666" [ 1014.176390] env[63372]: _type = "Task" [ 1014.176390] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.189787] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024666, 'name': Rename_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.277405] env[63372]: DEBUG oslo_concurrency.lockutils [req-cdb42f53-703b-415a-8a76-c95d3da1e709 req-22654d5d-bc20-45a6-9716-acc89f42aabc service nova] Releasing lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.330756] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024657, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.419117] env[63372]: DEBUG nova.objects.instance [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lazy-loading 'numa_topology' on Instance uuid 9be70e66-2d81-4c7b-8dcc-e69815faef1a {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.433173] env[63372]: DEBUG oslo_concurrency.lockutils [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.433389] env[63372]: DEBUG oslo_concurrency.lockutils [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.433568] env[63372]: DEBUG nova.network.neutron [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1014.433753] env[63372]: DEBUG nova.objects.instance [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lazy-loading 'info_cache' on Instance uuid cbb471c5-8d86-4d8f-8935-f6a77e5144be {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.463848] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024664, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.718462} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.464200] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 28eb1675-b1b8-46a6-873d-5c858b716575/28eb1675-b1b8-46a6-873d-5c858b716575.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1014.464454] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1014.464727] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa6c3751-ca19-4b34-a539-4756ad24302b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.473022] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1014.473022] env[63372]: value = "task-1024667" [ 1014.473022] env[63372]: _type = "Task" [ 1014.473022] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.479925] env[63372]: DEBUG oslo_vmware.api [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024665, 'name': ReconfigVM_Task, 'duration_secs': 0.226062} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.480620] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227445', 'volume_id': '2d01e3c9-8f3a-4399-a8a8-213b5e212d9a', 'name': 'volume-2d01e3c9-8f3a-4399-a8a8-213b5e212d9a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b7732621-7e58-40a3-b723-5c66df6f74a1', 'attached_at': '', 'detached_at': '', 'volume_id': '2d01e3c9-8f3a-4399-a8a8-213b5e212d9a', 'serial': '2d01e3c9-8f3a-4399-a8a8-213b5e212d9a'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1014.485623] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024667, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.590148] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.624981] env[63372]: DEBUG oslo_concurrency.lockutils [None req-932178f3-f48e-4d3e-a112-b18cace36df9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-b455b16b-3332-43bf-a91e-a350287ba5f3-1d17bf13-a03b-4654-bdfe-adc2d47e273a" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.995s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.686248] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024666, 'name': Rename_Task, 'duration_secs': 0.312281} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.686457] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1014.686695] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4db8a9b1-b319-4140-bca1-8cd9bf611b9c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.693126] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1014.693126] env[63372]: value = "task-1024668" [ 1014.693126] env[63372]: _type = "Task" [ 1014.693126] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.701905] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024668, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.829425] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024657, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.921869] env[63372]: DEBUG nova.objects.base [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Object Instance<9be70e66-2d81-4c7b-8dcc-e69815faef1a> lazy-loaded attributes: resources,numa_topology {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1014.968117] env[63372]: DEBUG nova.compute.manager [req-5ad7dbc9-b566-4fd4-85f9-23709c91c9bb req-0a231e2d-d4b0-46da-a705-57462d0ccf63 service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Received event network-vif-unplugged-9c53b112-6225-4e7e-95d5-cce3c462a615 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.968346] env[63372]: DEBUG oslo_concurrency.lockutils [req-5ad7dbc9-b566-4fd4-85f9-23709c91c9bb req-0a231e2d-d4b0-46da-a705-57462d0ccf63 service nova] Acquiring lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.968561] env[63372]: DEBUG oslo_concurrency.lockutils [req-5ad7dbc9-b566-4fd4-85f9-23709c91c9bb req-0a231e2d-d4b0-46da-a705-57462d0ccf63 service nova] Lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.968755] env[63372]: DEBUG oslo_concurrency.lockutils [req-5ad7dbc9-b566-4fd4-85f9-23709c91c9bb req-0a231e2d-d4b0-46da-a705-57462d0ccf63 service nova] Lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.968945] env[63372]: DEBUG nova.compute.manager [req-5ad7dbc9-b566-4fd4-85f9-23709c91c9bb req-0a231e2d-d4b0-46da-a705-57462d0ccf63 service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] No waiting events found dispatching network-vif-unplugged-9c53b112-6225-4e7e-95d5-cce3c462a615 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.969153] env[63372]: DEBUG nova.compute.manager [req-5ad7dbc9-b566-4fd4-85f9-23709c91c9bb req-0a231e2d-d4b0-46da-a705-57462d0ccf63 service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Received event network-vif-unplugged-9c53b112-6225-4e7e-95d5-cce3c462a615 for instance with task_state deleting. {{(pid=63372) _process_instance_event /opt/stack/nova/nova/compute/manager.py:10909}} [ 1014.969323] env[63372]: DEBUG nova.compute.manager [req-5ad7dbc9-b566-4fd4-85f9-23709c91c9bb req-0a231e2d-d4b0-46da-a705-57462d0ccf63 service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Received event network-changed-9c53b112-6225-4e7e-95d5-cce3c462a615 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.969532] env[63372]: DEBUG nova.compute.manager [req-5ad7dbc9-b566-4fd4-85f9-23709c91c9bb req-0a231e2d-d4b0-46da-a705-57462d0ccf63 service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Refreshing instance network info cache due to event network-changed-9c53b112-6225-4e7e-95d5-cce3c462a615. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1014.969738] env[63372]: DEBUG oslo_concurrency.lockutils [req-5ad7dbc9-b566-4fd4-85f9-23709c91c9bb req-0a231e2d-d4b0-46da-a705-57462d0ccf63 service nova] Acquiring lock "refresh_cache-9be70e66-2d81-4c7b-8dcc-e69815faef1a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.969858] env[63372]: DEBUG oslo_concurrency.lockutils [req-5ad7dbc9-b566-4fd4-85f9-23709c91c9bb req-0a231e2d-d4b0-46da-a705-57462d0ccf63 service nova] Acquired lock "refresh_cache-9be70e66-2d81-4c7b-8dcc-e69815faef1a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.970040] env[63372]: DEBUG nova.network.neutron [req-5ad7dbc9-b566-4fd4-85f9-23709c91c9bb req-0a231e2d-d4b0-46da-a705-57462d0ccf63 service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Refreshing network info cache for port 9c53b112-6225-4e7e-95d5-cce3c462a615 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1014.985717] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024667, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069686} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.988320] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1014.991316] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceecf1db-8b49-4ea1-8c9d-7bbd09e9d8be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.014307] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 28eb1675-b1b8-46a6-873d-5c858b716575/28eb1675-b1b8-46a6-873d-5c858b716575.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1015.017436] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-997136af-d1b0-4eb8-a6da-b99f750f08af {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.038833] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1015.038833] env[63372]: value = "task-1024669" [ 1015.038833] env[63372]: _type = "Task" [ 1015.038833] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.051213] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024669, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.120039] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1015.120311] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4cc3ed5-834d-4588-8008-6641a6a5f989 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.124667] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0282c3-487d-4fac-ba06-efb51af67605 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.129110] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1015.129110] env[63372]: value = "task-1024670" [ 1015.129110] env[63372]: _type = "Task" [ 1015.129110] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.136803] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a8e777-399a-455e-878a-cb9aeed5b4e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.143609] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024670, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.173819] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad78f3a8-bd0a-4beb-bc12-78825ea847b0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.181907] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e07af2e-47ef-4506-903b-812bfa3f8684 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.196656] env[63372]: DEBUG nova.compute.provider_tree [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.206119] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024668, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.288986] env[63372]: DEBUG oslo_concurrency.lockutils [None req-95b43c63-c882-4b72-9732-24654444a7b8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.329553] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024657, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.535731] env[63372]: DEBUG nova.objects.instance [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'flavor' on Instance uuid b7732621-7e58-40a3-b723-5c66df6f74a1 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.552914] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024669, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.646124] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024670, 'name': PowerOffVM_Task, 'duration_secs': 0.378742} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.646500] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1015.647617] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a3c81b7-f80e-43b4-8c10-5a53e6ba961b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.672737] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c544e26f-43f0-4463-b73a-da6687c74681 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.704479] env[63372]: DEBUG nova.scheduler.client.report [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1015.713995] env[63372]: DEBUG oslo_vmware.api [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024668, 'name': PowerOnVM_Task, 'duration_secs': 0.964892} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.716660] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1015.716660] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1015.716789] env[63372]: INFO nova.compute.manager [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Took 10.21 seconds to spawn the instance on the hypervisor. [ 1015.717047] env[63372]: DEBUG nova.compute.manager [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.717293] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a51ea50-4bd2-45a5-86e7-3778c6ae111c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.719704] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90612d3b-cdf8-4b80-b4cb-de5b2d950e6d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.732316] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1015.732316] env[63372]: value = "task-1024671" [ 1015.732316] env[63372]: _type = "Task" [ 1015.732316] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.743183] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] VM already powered off {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1015.743501] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1015.743833] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.744077] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.744329] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1015.744656] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-73e631db-ef70-4bdc-8c27-a4d971f76657 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.753484] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1015.753686] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1015.754453] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc819adf-9577-4a6d-aeac-36f6d417a019 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.760515] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1015.760515] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52603e5d-906d-9f3d-3cb2-83d5b59f9429" [ 1015.760515] env[63372]: _type = "Task" [ 1015.760515] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.768756] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52603e5d-906d-9f3d-3cb2-83d5b59f9429, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.830701] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024657, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.034922] env[63372]: DEBUG nova.network.neutron [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating instance_info_cache with network_info: [{"id": "298d217f-c17a-4ee2-acae-baee355112dd", "address": "fa:16:3e:66:af:86", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298d217f-c1", "ovs_interfaceid": "298d217f-c17a-4ee2-acae-baee355112dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.048025] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f2e65e0-f721-4dd0-a7dc-1ba2f1e664bc tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "b7732621-7e58-40a3-b723-5c66df6f74a1" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.369s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.065224] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "interface-b455b16b-3332-43bf-a91e-a350287ba5f3-1d17bf13-a03b-4654-bdfe-adc2d47e273a" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.065617] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-b455b16b-3332-43bf-a91e-a350287ba5f3-1d17bf13-a03b-4654-bdfe-adc2d47e273a" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.002s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.068465] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024669, 'name': ReconfigVM_Task, 'duration_secs': 0.582893} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.069630] env[63372]: DEBUG nova.network.neutron [req-5ad7dbc9-b566-4fd4-85f9-23709c91c9bb req-0a231e2d-d4b0-46da-a705-57462d0ccf63 service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Updated VIF entry in instance network info cache for port 9c53b112-6225-4e7e-95d5-cce3c462a615. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1016.070083] env[63372]: DEBUG nova.network.neutron [req-5ad7dbc9-b566-4fd4-85f9-23709c91c9bb req-0a231e2d-d4b0-46da-a705-57462d0ccf63 service nova] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Updating instance_info_cache with network_info: [{"id": "9c53b112-6225-4e7e-95d5-cce3c462a615", "address": "fa:16:3e:70:bf:65", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": null, "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9c53b112-62", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.073782] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 28eb1675-b1b8-46a6-873d-5c858b716575/28eb1675-b1b8-46a6-873d-5c858b716575.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.074576] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67deeab3-50ba-4753-9e23-2beb68058585 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.085744] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1016.085744] env[63372]: value = "task-1024672" [ 1016.085744] env[63372]: _type = "Task" [ 1016.085744] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.097580] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024672, 'name': Rename_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.216040] env[63372]: DEBUG oslo_concurrency.lockutils [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.306s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.243189] env[63372]: INFO nova.compute.manager [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Took 20.50 seconds to build instance. [ 1016.274856] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52603e5d-906d-9f3d-3cb2-83d5b59f9429, 'name': SearchDatastore_Task, 'duration_secs': 0.017754} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.276445] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05596816-7c85-46ac-9f5b-e4148f24780a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.284114] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1016.284114] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52a7f98c-609b-e187-2540-f4bda5bdd113" [ 1016.284114] env[63372]: _type = "Task" [ 1016.284114] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.296724] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a7f98c-609b-e187-2540-f4bda5bdd113, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.330832] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024657, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.538299] env[63372]: DEBUG oslo_concurrency.lockutils [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "refresh_cache-cbb471c5-8d86-4d8f-8935-f6a77e5144be" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.538580] env[63372]: DEBUG nova.objects.instance [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lazy-loading 'migration_context' on Instance uuid cbb471c5-8d86-4d8f-8935-f6a77e5144be {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.569395] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.569610] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.570482] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899fe817-da95-49df-8ef1-a27149e2c90e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.573433] env[63372]: DEBUG oslo_concurrency.lockutils [req-5ad7dbc9-b566-4fd4-85f9-23709c91c9bb req-0a231e2d-d4b0-46da-a705-57462d0ccf63 service nova] Releasing lock "refresh_cache-9be70e66-2d81-4c7b-8dcc-e69815faef1a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.592594] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daee6fc2-4fea-4e30-bcd1-1024dba20643 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.600260] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024672, 'name': Rename_Task, 'duration_secs': 0.20426} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.615838] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1016.621918] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Reconfiguring VM to detach interface {{(pid=63372) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1016.622418] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01edd28f-a25c-4584-83d8-7a5f0e0a2975 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.624410] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e16acd8-52f6-4791-b488-1c1b97c44746 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.643145] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1016.643145] env[63372]: value = "task-1024673" [ 1016.643145] env[63372]: _type = "Task" [ 1016.643145] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.644528] env[63372]: DEBUG oslo_vmware.api [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 1016.644528] env[63372]: value = "task-1024674" [ 1016.644528] env[63372]: _type = "Task" [ 1016.644528] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.655747] env[63372]: DEBUG oslo_vmware.api [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024674, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.658312] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024673, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.725152] env[63372]: DEBUG oslo_concurrency.lockutils [None req-59ad2a4c-03ff-453e-a949-91bf9996e2d4 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.394s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.726135] env[63372]: DEBUG oslo_concurrency.lockutils [None req-95b43c63-c882-4b72-9732-24654444a7b8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 1.437s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.726506] env[63372]: DEBUG oslo_concurrency.lockutils [None req-95b43c63-c882-4b72-9732-24654444a7b8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.726609] env[63372]: DEBUG oslo_concurrency.lockutils [None req-95b43c63-c882-4b72-9732-24654444a7b8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.726733] env[63372]: DEBUG oslo_concurrency.lockutils [None req-95b43c63-c882-4b72-9732-24654444a7b8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.728452] env[63372]: INFO nova.compute.manager [None req-95b43c63-c882-4b72-9732-24654444a7b8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Terminating instance [ 1016.730230] env[63372]: DEBUG nova.compute.manager [None req-95b43c63-c882-4b72-9732-24654444a7b8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1016.730427] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-95b43c63-c882-4b72-9732-24654444a7b8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1016.730880] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91b7852a-a0e6-41e2-a7eb-ca4a7160c149 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.740029] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8823fa0c-3b9f-43f7-ac4f-6276321c1eb6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.751501] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c52e1460-1bd9-45d7-9277-e97c6634ecc4 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "4a2e53e0-b68c-42be-afff-1eead017679d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.025s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.773157] env[63372]: WARNING nova.virt.vmwareapi.vmops [None req-95b43c63-c882-4b72-9732-24654444a7b8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9be70e66-2d81-4c7b-8dcc-e69815faef1a could not be found. [ 1016.773395] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-95b43c63-c882-4b72-9732-24654444a7b8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1016.773581] env[63372]: INFO nova.compute.manager [None req-95b43c63-c882-4b72-9732-24654444a7b8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1016.773842] env[63372]: DEBUG oslo.service.loopingcall [None req-95b43c63-c882-4b72-9732-24654444a7b8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.774112] env[63372]: DEBUG nova.compute.manager [-] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1016.774208] env[63372]: DEBUG nova.network.neutron [-] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1016.795839] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52a7f98c-609b-e187-2540-f4bda5bdd113, 'name': SearchDatastore_Task, 'duration_secs': 0.02306} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.796143] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.796409] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] ddca7aa5-fcc8-4835-bfdb-47781335f2cd/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk. {{(pid=63372) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1016.796677] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b09208b0-5f2c-424e-bf9b-caaa530550d4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.803449] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1016.803449] env[63372]: value = "task-1024675" [ 1016.803449] env[63372]: _type = "Task" [ 1016.803449] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.813471] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024675, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.830761] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024657, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.021021] env[63372]: DEBUG nova.compute.manager [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Stashing vm_state: active {{(pid=63372) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1017.042610] env[63372]: DEBUG nova.objects.base [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1017.043042] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31981004-6a5c-4fec-8ba1-71b8a8de1f72 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.074955] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c1b8587-5680-4320-8751-0f06d7fc1b40 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.081934] env[63372]: DEBUG oslo_vmware.api [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1017.081934] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525ab130-2408-e425-9f55-a93dc78f466f" [ 1017.081934] env[63372]: _type = "Task" [ 1017.081934] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.091984] env[63372]: DEBUG oslo_vmware.api [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525ab130-2408-e425-9f55-a93dc78f466f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.161273] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024673, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.165365] env[63372]: DEBUG oslo_vmware.api [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.314093] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024675, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.332499] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024657, 'name': CreateVM_Task, 'duration_secs': 4.709709} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.332675] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1017.333414] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.333582] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.333945] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1017.334247] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99dca16f-f7d5-4fd8-96ca-0a500bb1de72 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.339306] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 1017.339306] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523aad26-b52a-010e-a325-3502b6ef8958" [ 1017.339306] env[63372]: _type = "Task" [ 1017.339306] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.348439] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523aad26-b52a-010e-a325-3502b6ef8958, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.491480] env[63372]: DEBUG nova.network.neutron [-] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.525700] env[63372]: DEBUG nova.compute.manager [req-b57f5ba2-0370-442e-8e60-3c6b40ddd088 req-18b2ce98-ce8e-46a7-9aff-8828c258b603 service nova] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Received event network-changed-7cd17ebc-3698-499b-8a5a-0507f100485c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.525700] env[63372]: DEBUG nova.compute.manager [req-b57f5ba2-0370-442e-8e60-3c6b40ddd088 req-18b2ce98-ce8e-46a7-9aff-8828c258b603 service nova] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Refreshing instance network info cache due to event network-changed-7cd17ebc-3698-499b-8a5a-0507f100485c. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1017.525700] env[63372]: DEBUG oslo_concurrency.lockutils [req-b57f5ba2-0370-442e-8e60-3c6b40ddd088 req-18b2ce98-ce8e-46a7-9aff-8828c258b603 service nova] Acquiring lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.525700] env[63372]: DEBUG oslo_concurrency.lockutils [req-b57f5ba2-0370-442e-8e60-3c6b40ddd088 req-18b2ce98-ce8e-46a7-9aff-8828c258b603 service nova] Acquired lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.525700] env[63372]: DEBUG nova.network.neutron [req-b57f5ba2-0370-442e-8e60-3c6b40ddd088 req-18b2ce98-ce8e-46a7-9aff-8828c258b603 service nova] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Refreshing network info cache for port 7cd17ebc-3698-499b-8a5a-0507f100485c {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1017.542834] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.543152] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.594736] env[63372]: DEBUG oslo_vmware.api [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525ab130-2408-e425-9f55-a93dc78f466f, 'name': SearchDatastore_Task, 'duration_secs': 0.012744} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.595446] env[63372]: DEBUG oslo_concurrency.lockutils [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.660178] env[63372]: DEBUG oslo_vmware.api [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024673, 'name': PowerOnVM_Task, 'duration_secs': 0.728186} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.666099] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1017.666099] env[63372]: INFO nova.compute.manager [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Took 9.78 seconds to spawn the instance on the hypervisor. [ 1017.666099] env[63372]: DEBUG nova.compute.manager [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1017.666099] env[63372]: DEBUG oslo_vmware.api [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.666099] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66539d7b-c199-49cf-a6f7-7fd0bc229c4f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.814042] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024675, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547164} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.814322] env[63372]: INFO nova.virt.vmwareapi.ds_util [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] ddca7aa5-fcc8-4835-bfdb-47781335f2cd/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk. [ 1017.815081] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc3dd52-6cc7-4755-b17e-91f688026632 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.839732] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] ddca7aa5-fcc8-4835-bfdb-47781335f2cd/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.840117] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-79d56ce0-3fad-499f-9269-bea6e336a189 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.861072] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523aad26-b52a-010e-a325-3502b6ef8958, 'name': SearchDatastore_Task, 'duration_secs': 0.02478} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.862208] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.862447] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1017.862700] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.862850] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.863043] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1017.863341] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1017.863341] env[63372]: value = "task-1024676" [ 1017.863341] env[63372]: _type = "Task" [ 1017.863341] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.863516] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7644f19-1515-400f-923a-9e2a3a8a26f2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.872711] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024676, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.878649] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1017.878830] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1017.879543] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac04af1d-9eb4-4db3-9ef3-f85f498dd9a4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.884239] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 1017.884239] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52901aeb-a939-58b4-49f0-f2d068132640" [ 1017.884239] env[63372]: _type = "Task" [ 1017.884239] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.891798] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52901aeb-a939-58b4-49f0-f2d068132640, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.994579] env[63372]: INFO nova.compute.manager [-] [instance: 9be70e66-2d81-4c7b-8dcc-e69815faef1a] Took 1.22 seconds to deallocate network for instance. [ 1018.048431] env[63372]: INFO nova.compute.claims [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1018.162079] env[63372]: DEBUG oslo_vmware.api [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.188222] env[63372]: INFO nova.compute.manager [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Took 20.68 seconds to build instance. [ 1018.259310] env[63372]: DEBUG nova.network.neutron [req-b57f5ba2-0370-442e-8e60-3c6b40ddd088 req-18b2ce98-ce8e-46a7-9aff-8828c258b603 service nova] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updated VIF entry in instance network info cache for port 7cd17ebc-3698-499b-8a5a-0507f100485c. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1018.259665] env[63372]: DEBUG nova.network.neutron [req-b57f5ba2-0370-442e-8e60-3c6b40ddd088 req-18b2ce98-ce8e-46a7-9aff-8828c258b603 service nova] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance_info_cache with network_info: [{"id": "7cd17ebc-3698-499b-8a5a-0507f100485c", "address": "fa:16:3e:59:5c:0e", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cd17ebc-36", "ovs_interfaceid": "7cd17ebc-3698-499b-8a5a-0507f100485c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.379061] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024676, 'name': ReconfigVM_Task, 'duration_secs': 0.323571} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.379415] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Reconfigured VM instance instance-00000060 to attach disk [datastore2] ddca7aa5-fcc8-4835-bfdb-47781335f2cd/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.380318] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66016c67-09f7-45d1-a3f9-7adc1a59dc33 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.410506] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e959ba89-5ce2-4b9b-b155-f56c8a9b4e44 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.424857] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52901aeb-a939-58b4-49f0-f2d068132640, 'name': SearchDatastore_Task, 'duration_secs': 0.023345} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.425996] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61b31479-0713-478d-8a0b-8efebb40d0f7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.430342] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1018.430342] env[63372]: value = "task-1024677" [ 1018.430342] env[63372]: _type = "Task" [ 1018.430342] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.431798] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 1018.431798] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]526a241c-a485-bdf5-7351-22c2eca42947" [ 1018.431798] env[63372]: _type = "Task" [ 1018.431798] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.442926] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526a241c-a485-bdf5-7351-22c2eca42947, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.446493] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024677, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.557336] env[63372]: INFO nova.compute.resource_tracker [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating resource usage from migration 7acff7e3-4df1-4ed7-8106-7ee02c83b666 [ 1018.661885] env[63372]: DEBUG oslo_vmware.api [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.690573] env[63372]: DEBUG oslo_concurrency.lockutils [None req-64da57c6-7b5c-45c1-bd22-6673de9762cc tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "28eb1675-b1b8-46a6-873d-5c858b716575" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.204s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.729346] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964ccbba-cd3a-4054-8500-221b6edc224c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.737108] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c387a5f-912b-4f8a-a23c-751213e684ba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.767283] env[63372]: DEBUG oslo_concurrency.lockutils [req-b57f5ba2-0370-442e-8e60-3c6b40ddd088 req-18b2ce98-ce8e-46a7-9aff-8828c258b603 service nova] Releasing lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.768764] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7696dd-f966-4854-bbe1-283e6db2216c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.776670] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1c8914-d2ce-4e7a-9187-5336d07e4c9d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.791727] env[63372]: DEBUG nova.compute.provider_tree [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.945859] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526a241c-a485-bdf5-7351-22c2eca42947, 'name': SearchDatastore_Task, 'duration_secs': 0.020134} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.949021] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.949378] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 800cb48b-e75e-45e7-94dc-6ad147e6c6c5/800cb48b-e75e-45e7-94dc-6ad147e6c6c5.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1018.949704] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024677, 'name': ReconfigVM_Task, 'duration_secs': 0.152006} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.949914] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f83cf2ae-74c1-4373-8ee1-22fae282562e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.952083] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1018.952392] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-994fab74-627f-4472-b2fe-677453ec5b89 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.959033] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 1018.959033] env[63372]: value = "task-1024678" [ 1018.959033] env[63372]: _type = "Task" [ 1018.959033] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.960347] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1018.960347] env[63372]: value = "task-1024679" [ 1018.960347] env[63372]: _type = "Task" [ 1018.960347] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.970728] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024678, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.973686] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024679, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.021302] env[63372]: DEBUG oslo_concurrency.lockutils [None req-95b43c63-c882-4b72-9732-24654444a7b8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "9be70e66-2d81-4c7b-8dcc-e69815faef1a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.295s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.162565] env[63372]: DEBUG oslo_vmware.api [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.298025] env[63372]: DEBUG nova.scheduler.client.report [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.477429] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024678, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.482045] env[63372]: DEBUG oslo_vmware.api [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024679, 'name': PowerOnVM_Task, 'duration_secs': 0.382882} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.484084] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1019.491039] env[63372]: DEBUG nova.compute.manager [None req-ddd78fba-35bf-43c9-80e6-bc7a7135bf5a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1019.491039] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723697ad-13fa-432f-b23d-8c35e6a0abb7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.663707] env[63372]: DEBUG oslo_vmware.api [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.804020] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.259s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.804020] env[63372]: INFO nova.compute.manager [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Migrating [ 1019.812888] env[63372]: DEBUG oslo_concurrency.lockutils [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.218s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.974353] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024678, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.619335} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.974353] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 800cb48b-e75e-45e7-94dc-6ad147e6c6c5/800cb48b-e75e-45e7-94dc-6ad147e6c6c5.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1019.974353] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1019.974353] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ceaabe75-d69b-4d7b-95e9-6b731c4bb99e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.980347] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 1019.980347] env[63372]: value = "task-1024680" [ 1019.980347] env[63372]: _type = "Task" [ 1019.980347] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.990831] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024680, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.013395] env[63372]: DEBUG nova.compute.manager [req-dd18d952-4e95-4f13-91f7-383cde384bd0 req-a5640128-4098-4e09-a3be-449ffc85797d service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Received event network-changed-845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.013609] env[63372]: DEBUG nova.compute.manager [req-dd18d952-4e95-4f13-91f7-383cde384bd0 req-a5640128-4098-4e09-a3be-449ffc85797d service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Refreshing instance network info cache due to event network-changed-845b30fa-3e7e-4324-9479-aed7aade3f91. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1020.013823] env[63372]: DEBUG oslo_concurrency.lockutils [req-dd18d952-4e95-4f13-91f7-383cde384bd0 req-a5640128-4098-4e09-a3be-449ffc85797d service nova] Acquiring lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.013977] env[63372]: DEBUG oslo_concurrency.lockutils [req-dd18d952-4e95-4f13-91f7-383cde384bd0 req-a5640128-4098-4e09-a3be-449ffc85797d service nova] Acquired lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.014798] env[63372]: DEBUG nova.network.neutron [req-dd18d952-4e95-4f13-91f7-383cde384bd0 req-a5640128-4098-4e09-a3be-449ffc85797d service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Refreshing network info cache for port 845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1020.166273] env[63372]: DEBUG oslo_vmware.api [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.322642] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.322875] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.323176] env[63372]: DEBUG nova.network.neutron [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1020.492636] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024680, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.165586} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.493337] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1020.493738] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77433127-3040-4bfe-b97d-b506aa0cfff5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.519134] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 800cb48b-e75e-45e7-94dc-6ad147e6c6c5/800cb48b-e75e-45e7-94dc-6ad147e6c6c5.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.523092] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8521c80-8529-4322-9b99-a990544cad98 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.539914] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74da96a6-6082-4e41-9951-141fc5529f33 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.551025] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a49c4dd-31c3-4974-a1ce-71c9b6fc22cf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.553401] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 1020.553401] env[63372]: value = "task-1024681" [ 1020.553401] env[63372]: _type = "Task" [ 1020.553401] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.591095] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92564b7-d043-4c1b-968c-6a132a9fe12a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.598012] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024681, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.603376] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f33b84-27af-48ad-bfec-03fca1dd4845 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.618523] env[63372]: DEBUG nova.compute.provider_tree [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.663389] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "c38256cc-215c-4bec-94e5-24a1bbef3e10" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.663655] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c38256cc-215c-4bec-94e5-24a1bbef3e10" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.671380] env[63372]: DEBUG oslo_vmware.api [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.799917] env[63372]: INFO nova.compute.manager [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Unrescuing [ 1020.800185] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.800372] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.800576] env[63372]: DEBUG nova.network.neutron [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1020.849915] env[63372]: DEBUG nova.network.neutron [req-dd18d952-4e95-4f13-91f7-383cde384bd0 req-a5640128-4098-4e09-a3be-449ffc85797d service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Updated VIF entry in instance network info cache for port 845b30fa-3e7e-4324-9479-aed7aade3f91. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1020.850368] env[63372]: DEBUG nova.network.neutron [req-dd18d952-4e95-4f13-91f7-383cde384bd0 req-a5640128-4098-4e09-a3be-449ffc85797d service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Updating instance_info_cache with network_info: [{"id": "845b30fa-3e7e-4324-9479-aed7aade3f91", "address": "fa:16:3e:54:d5:15", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap845b30fa-3e", "ovs_interfaceid": "845b30fa-3e7e-4324-9479-aed7aade3f91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.041041] env[63372]: DEBUG nova.network.neutron [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance_info_cache with network_info: [{"id": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "address": "fa:16:3e:28:02:b8", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap128a0686-7c", "ovs_interfaceid": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.064933] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024681, 'name': ReconfigVM_Task, 'duration_secs': 0.29908} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.065249] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 800cb48b-e75e-45e7-94dc-6ad147e6c6c5/800cb48b-e75e-45e7-94dc-6ad147e6c6c5.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.066092] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-350ffa54-c39e-4337-b941-c976f7dc92d4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.072328] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 1021.072328] env[63372]: value = "task-1024682" [ 1021.072328] env[63372]: _type = "Task" [ 1021.072328] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.079615] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024682, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.125046] env[63372]: DEBUG nova.scheduler.client.report [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1021.167106] env[63372]: DEBUG nova.compute.manager [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1021.169726] env[63372]: DEBUG oslo_vmware.api [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.353826] env[63372]: DEBUG oslo_concurrency.lockutils [req-dd18d952-4e95-4f13-91f7-383cde384bd0 req-a5640128-4098-4e09-a3be-449ffc85797d service nova] Releasing lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.512469] env[63372]: DEBUG nova.network.neutron [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Updating instance_info_cache with network_info: [{"id": "b54d2aaa-a37b-4db7-b735-f9d2a6a9534b", "address": "fa:16:3e:03:b6:cc", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb54d2aaa-a3", "ovs_interfaceid": "b54d2aaa-a37b-4db7-b735-f9d2a6a9534b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.543960] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.582586] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024682, 'name': Rename_Task, 'duration_secs': 0.147211} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.582923] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1021.583272] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ce8d1a7-488d-4bda-b7d9-4c8e81009c9b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.590214] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 1021.590214] env[63372]: value = "task-1024683" [ 1021.590214] env[63372]: _type = "Task" [ 1021.590214] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.598935] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024683, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.667564] env[63372]: DEBUG oslo_vmware.api [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.691532] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524910d3-3651-60c9-5d0a-4207788274da/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1021.692761] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b82b94bc-7ce4-4aef-8815-453f0accf207 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.696581] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.700765] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524910d3-3651-60c9-5d0a-4207788274da/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1021.701080] env[63372]: ERROR oslo_vmware.rw_handles [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524910d3-3651-60c9-5d0a-4207788274da/disk-0.vmdk due to incomplete transfer. [ 1021.701424] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-27253b7d-0765-41c6-8b7e-4fcbea12f293 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.709300] env[63372]: DEBUG oslo_vmware.rw_handles [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524910d3-3651-60c9-5d0a-4207788274da/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1021.709508] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Uploaded image 003d2e97-266b-4459-abd7-6ca82ed4825e to the Glance image server {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1021.711407] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1021.711673] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-14351553-5c02-48fa-bb15-9d3c1af5b68c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.718052] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1021.718052] env[63372]: value = "task-1024684" [ 1021.718052] env[63372]: _type = "Task" [ 1021.718052] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.726802] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024684, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.015941] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.016592] env[63372]: DEBUG nova.objects.instance [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lazy-loading 'flavor' on Instance uuid ddca7aa5-fcc8-4835-bfdb-47781335f2cd {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.101871] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024683, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.136388] env[63372]: DEBUG oslo_concurrency.lockutils [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.323s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.141808] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.444s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.143461] env[63372]: INFO nova.compute.claims [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1022.167669] env[63372]: DEBUG oslo_vmware.api [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024674, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.229694] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024684, 'name': Destroy_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.522700] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1757ead0-ad0c-465b-9d5e-7ee2ffdd55c5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.543900] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1022.544307] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e9ce1bc-a3fa-4a17-819b-74959602a4af {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.553115] env[63372]: DEBUG oslo_vmware.api [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1022.553115] env[63372]: value = "task-1024685" [ 1022.553115] env[63372]: _type = "Task" [ 1022.553115] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.562427] env[63372]: DEBUG oslo_vmware.api [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024685, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.599915] env[63372]: DEBUG oslo_vmware.api [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024683, 'name': PowerOnVM_Task, 'duration_secs': 0.514691} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.600195] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1022.600402] env[63372]: INFO nova.compute.manager [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Took 12.41 seconds to spawn the instance on the hypervisor. [ 1022.600582] env[63372]: DEBUG nova.compute.manager [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.601438] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04a3842-80fd-4246-8926-bfeb0dff81fc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.668971] env[63372]: DEBUG oslo_vmware.api [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024674, 'name': ReconfigVM_Task, 'duration_secs': 5.939565} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.669250] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.669455] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Reconfigured VM to detach interface {{(pid=63372) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1022.698215] env[63372]: INFO nova.scheduler.client.report [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleted allocation for migration 75fa60a2-d2ae-4672-8d3c-8a09a4cdd49a [ 1022.729970] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024684, 'name': Destroy_Task, 'duration_secs': 0.858741} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.729970] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Destroyed the VM [ 1022.730205] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1022.730332] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d772003c-18cf-40c7-91f5-69bae107a39e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.736708] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1022.736708] env[63372]: value = "task-1024686" [ 1022.736708] env[63372]: _type = "Task" [ 1022.736708] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.746239] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024686, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.060673] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf74c019-457b-4be9-b18a-e41a515cbb5f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.068599] env[63372]: DEBUG oslo_vmware.api [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024685, 'name': PowerOffVM_Task, 'duration_secs': 0.220329} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.085610] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1023.090983] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Reconfiguring VM instance instance-00000060 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1023.091723] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance 'b7732621-7e58-40a3-b723-5c66df6f74a1' progress to 0 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1023.095066] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8cc8e88-45af-497f-a2d1-a78954a9ba89 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.115692] env[63372]: DEBUG oslo_vmware.api [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1023.115692] env[63372]: value = "task-1024687" [ 1023.115692] env[63372]: _type = "Task" [ 1023.115692] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.121236] env[63372]: INFO nova.compute.manager [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Took 22.04 seconds to build instance. [ 1023.127411] env[63372]: DEBUG oslo_vmware.api [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024687, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.205882] env[63372]: DEBUG oslo_concurrency.lockutils [None req-83800660-23b4-4553-a360-10e29bae939e tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.352s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.246890] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024686, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.338132] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-536263a8-1ef3-4263-821c-b04cff41b4d1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.344446] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6b9f1ae-d72d-4741-90c3-ff500447317f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.380492] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed08d1f3-443c-4c37-8cba-e45de1902adb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.388044] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7d45be-1b8e-4c21-b573-1088815f2699 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.402045] env[63372]: DEBUG nova.compute.provider_tree [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.610866] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1023.611242] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6692423e-ca70-42a2-a041-eb8c867cc1db {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.627329] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fd65dca1-8d64-438a-818c-c617504f7512 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.560s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.627329] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1023.627329] env[63372]: value = "task-1024688" [ 1023.627329] env[63372]: _type = "Task" [ 1023.627329] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.634748] env[63372]: DEBUG oslo_vmware.api [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024687, 'name': ReconfigVM_Task, 'duration_secs': 0.391562} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.635499] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Reconfigured VM instance instance-00000060 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1023.635708] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1023.635961] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6914efdf-ebda-4397-86db-9ecf4c092216 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.641606] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024688, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.645958] env[63372]: DEBUG oslo_vmware.api [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1023.645958] env[63372]: value = "task-1024689" [ 1023.645958] env[63372]: _type = "Task" [ 1023.645958] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.654108] env[63372]: DEBUG oslo_vmware.api [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024689, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.748090] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024686, 'name': RemoveSnapshot_Task, 'duration_secs': 0.588539} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.749588] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1023.753032] env[63372]: DEBUG nova.compute.manager [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1023.753032] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378f888c-e5f5-447a-a9b3-ae70c7f4e22f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.907077] env[63372]: DEBUG nova.scheduler.client.report [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.000265] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.000998] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.002754] env[63372]: DEBUG nova.network.neutron [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1024.137610] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024688, 'name': PowerOffVM_Task, 'duration_secs': 0.26511} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.137961] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1024.138072] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance 'b7732621-7e58-40a3-b723-5c66df6f74a1' progress to 17 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1024.156101] env[63372]: DEBUG oslo_vmware.api [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024689, 'name': PowerOnVM_Task, 'duration_secs': 0.393248} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.156465] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1024.156856] env[63372]: DEBUG nova.compute.manager [None req-c89545c4-6138-48be-bd5b-ebdc46eff28a tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.158144] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5d522a-454b-4e95-9e98-3f9f1ae835d9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.265204] env[63372]: INFO nova.compute.manager [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Shelve offloading [ 1024.266711] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1024.266959] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f724ac57-0333-4e19-a0f9-4eb72d0e7e84 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.274174] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1024.274174] env[63372]: value = "task-1024690" [ 1024.274174] env[63372]: _type = "Task" [ 1024.274174] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.283808] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] VM already powered off {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1024.283995] env[63372]: DEBUG nova.compute.manager [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.284843] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e35448-c1ca-4e8e-b3b2-1f5386349b35 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.290370] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.290536] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquired lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.290720] env[63372]: DEBUG nova.network.neutron [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1024.411588] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.411588] env[63372]: DEBUG nova.compute.manager [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1024.616245] env[63372]: DEBUG oslo_concurrency.lockutils [None req-519f0d34-90f7-439e-a352-7d892648b668 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.616588] env[63372]: DEBUG oslo_concurrency.lockutils [None req-519f0d34-90f7-439e-a352-7d892648b668 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.616802] env[63372]: DEBUG nova.compute.manager [None req-519f0d34-90f7-439e-a352-7d892648b668 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.618179] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8342590c-27b3-4de4-9881-3cefb6bdcf26 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.627112] env[63372]: DEBUG nova.compute.manager [None req-519f0d34-90f7-439e-a352-7d892648b668 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63372) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1024.627671] env[63372]: DEBUG nova.objects.instance [None req-519f0d34-90f7-439e-a352-7d892648b668 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lazy-loading 'flavor' on Instance uuid 800cb48b-e75e-45e7-94dc-6ad147e6c6c5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.644628] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1024.644881] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1024.645049] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1024.645240] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1024.645388] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1024.645569] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1024.645825] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1024.646064] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1024.646291] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1024.646485] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1024.646711] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1024.652591] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4845cab7-9959-4cb8-a4e0-14cd804a4241 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.671447] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1024.671447] env[63372]: value = "task-1024691" [ 1024.671447] env[63372]: _type = "Task" [ 1024.671447] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.682803] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024691, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.918922] env[63372]: DEBUG nova.compute.utils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1024.920358] env[63372]: DEBUG nova.compute.manager [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1024.920834] env[63372]: DEBUG nova.network.neutron [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1024.950079] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.950373] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.950600] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.950799] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.951058] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.953351] env[63372]: INFO nova.compute.manager [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Terminating instance [ 1024.957180] env[63372]: DEBUG nova.compute.manager [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1024.957353] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1024.958482] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bbf18b0-533e-4f34-b12e-08753bba112d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.962312] env[63372]: DEBUG nova.policy [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '118d2f1de2be45cc8bb48bb75525d37e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'afe39cfedf214d50be775ef736f94da9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1024.964146] env[63372]: INFO nova.network.neutron [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Port 1d17bf13-a03b-4654-bdfe-adc2d47e273a from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1024.964485] env[63372]: DEBUG nova.network.neutron [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updating instance_info_cache with network_info: [{"id": "18c7850a-d97e-4564-92e5-38d1204369e4", "address": "fa:16:3e:14:28:4a", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c7850a-d9", "ovs_interfaceid": "18c7850a-d97e-4564-92e5-38d1204369e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.970713] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1024.970987] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3af68c51-b99a-4f3f-93c0-7852c6b9b21d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.978050] env[63372]: DEBUG oslo_vmware.api [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1024.978050] env[63372]: value = "task-1024692" [ 1024.978050] env[63372]: _type = "Task" [ 1024.978050] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.990824] env[63372]: DEBUG oslo_vmware.api [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024692, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.133153] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-519f0d34-90f7-439e-a352-7d892648b668 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1025.133442] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9bc1b1e1-3797-479b-980d-73036142a226 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.141009] env[63372]: DEBUG oslo_vmware.api [None req-519f0d34-90f7-439e-a352-7d892648b668 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 1025.141009] env[63372]: value = "task-1024693" [ 1025.141009] env[63372]: _type = "Task" [ 1025.141009] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.149879] env[63372]: DEBUG oslo_vmware.api [None req-519f0d34-90f7-439e-a352-7d892648b668 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024693, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.183127] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024691, 'name': ReconfigVM_Task, 'duration_secs': 0.230642} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.183570] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance 'b7732621-7e58-40a3-b723-5c66df6f74a1' progress to 33 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1025.257326] env[63372]: DEBUG nova.network.neutron [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updating instance_info_cache with network_info: [{"id": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "address": "fa:16:3e:1f:74:b1", "network": {"id": "cbdcb945-f0a5-49b1-9b18-fcd8e4ca9728", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1068757189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e517915d5e7a4e3e924cc95c55170a66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57943c40-a8", "ovs_interfaceid": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.425338] env[63372]: DEBUG nova.compute.manager [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1025.430903] env[63372]: DEBUG nova.network.neutron [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Successfully created port: 9281c23e-1f88-4862-8f9a-e3e3478592dc {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1025.467973] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.476128] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.476128] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.493084] env[63372]: DEBUG oslo_vmware.api [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024692, 'name': PowerOffVM_Task, 'duration_secs': 0.410054} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.493604] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1025.493776] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1025.494040] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2549c2e-18eb-4b93-8abb-a3d5410fb7ee {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.656269] env[63372]: DEBUG oslo_vmware.api [None req-519f0d34-90f7-439e-a352-7d892648b668 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024693, 'name': PowerOffVM_Task, 'duration_secs': 0.208426} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.656269] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-519f0d34-90f7-439e-a352-7d892648b668 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1025.656269] env[63372]: DEBUG nova.compute.manager [None req-519f0d34-90f7-439e-a352-7d892648b668 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1025.656919] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b7b940c-5c0e-433b-a034-44c9c2b050d5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.681552] env[63372]: DEBUG nova.compute.manager [req-9b0a6a85-f3ce-49ca-b9fa-9291e165b974 req-37cc19ba-ad8c-450c-8d29-c2474d3445f5 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Received event network-changed-18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1025.681803] env[63372]: DEBUG nova.compute.manager [req-9b0a6a85-f3ce-49ca-b9fa-9291e165b974 req-37cc19ba-ad8c-450c-8d29-c2474d3445f5 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Refreshing instance network info cache due to event network-changed-18c7850a-d97e-4564-92e5-38d1204369e4. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1025.681918] env[63372]: DEBUG oslo_concurrency.lockutils [req-9b0a6a85-f3ce-49ca-b9fa-9291e165b974 req-37cc19ba-ad8c-450c-8d29-c2474d3445f5 service nova] Acquiring lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.682188] env[63372]: DEBUG oslo_concurrency.lockutils [req-9b0a6a85-f3ce-49ca-b9fa-9291e165b974 req-37cc19ba-ad8c-450c-8d29-c2474d3445f5 service nova] Acquired lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.683580] env[63372]: DEBUG nova.network.neutron [req-9b0a6a85-f3ce-49ca-b9fa-9291e165b974 req-37cc19ba-ad8c-450c-8d29-c2474d3445f5 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Refreshing network info cache for port 18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1025.693959] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1025.695231] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1025.696028] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1025.696028] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1025.696149] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1025.696424] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1025.696975] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1025.696975] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1025.697257] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1025.697540] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1025.698837] env[63372]: DEBUG nova.virt.hardware [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1025.709330] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Reconfiguring VM instance instance-0000005a to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1025.711054] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0cde9de-9c85-4778-94a1-7bf2bd44a071 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.734447] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1025.734447] env[63372]: value = "task-1024695" [ 1025.734447] env[63372]: _type = "Task" [ 1025.734447] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.746151] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024695, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.759730] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Releasing lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.948875] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1025.948875] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1025.948875] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleting the datastore file [datastore1] cbb471c5-8d86-4d8f-8935-f6a77e5144be {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.948875] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1deed084-91e8-4b23-9df7-9a40b8ef9eab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.955674] env[63372]: DEBUG oslo_vmware.api [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1025.955674] env[63372]: value = "task-1024696" [ 1025.955674] env[63372]: _type = "Task" [ 1025.955674] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.964231] env[63372]: DEBUG oslo_vmware.api [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024696, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.976365] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2e1947ce-2735-4979-bdc2-c3c526f1b4b9 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-b455b16b-3332-43bf-a91e-a350287ba5f3-1d17bf13-a03b-4654-bdfe-adc2d47e273a" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.911s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.979031] env[63372]: DEBUG nova.compute.manager [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1026.033516] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1026.034424] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e18eebb-4124-4c95-870d-d1217ec21218 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.042271] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1026.042511] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2628b96-366c-436c-a5a8-93e1807745cd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.099948] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1026.100184] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1026.100369] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Deleting the datastore file [datastore1] 91985614-b959-401e-bb06-d67b230ee026 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1026.100639] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3d84c48b-a988-4878-a330-ab5676bc5a32 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.108406] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1026.108406] env[63372]: value = "task-1024698" [ 1026.108406] env[63372]: _type = "Task" [ 1026.108406] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.117022] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024698, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.176367] env[63372]: DEBUG oslo_concurrency.lockutils [None req-519f0d34-90f7-439e-a352-7d892648b668 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.560s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.248206] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024695, 'name': ReconfigVM_Task, 'duration_secs': 0.264072} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.248603] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Reconfigured VM instance instance-0000005a to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1026.249625] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba6db96-ccf9-4fa9-b171-50001a3312b6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.269494] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "interface-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-1d17bf13-a03b-4654-bdfe-adc2d47e273a" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.269777] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-1d17bf13-a03b-4654-bdfe-adc2d47e273a" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.270389] env[63372]: DEBUG nova.objects.instance [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'flavor' on Instance uuid d7b7ee35-5e83-4c62-bd1e-8ec39951c44c {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.280058] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] b7732621-7e58-40a3-b723-5c66df6f74a1/b7732621-7e58-40a3-b723-5c66df6f74a1.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1026.283120] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a57922cb-2f32-4b47-9a2a-696944d534fd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.303608] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1026.303608] env[63372]: value = "task-1024699" [ 1026.303608] env[63372]: _type = "Task" [ 1026.303608] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.314386] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024699, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.441942] env[63372]: DEBUG nova.compute.manager [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1026.444801] env[63372]: DEBUG nova.network.neutron [req-9b0a6a85-f3ce-49ca-b9fa-9291e165b974 req-37cc19ba-ad8c-450c-8d29-c2474d3445f5 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updated VIF entry in instance network info cache for port 18c7850a-d97e-4564-92e5-38d1204369e4. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1026.445158] env[63372]: DEBUG nova.network.neutron [req-9b0a6a85-f3ce-49ca-b9fa-9291e165b974 req-37cc19ba-ad8c-450c-8d29-c2474d3445f5 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updating instance_info_cache with network_info: [{"id": "18c7850a-d97e-4564-92e5-38d1204369e4", "address": "fa:16:3e:14:28:4a", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c7850a-d9", "ovs_interfaceid": "18c7850a-d97e-4564-92e5-38d1204369e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.466649] env[63372]: DEBUG oslo_vmware.api [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024696, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154392} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.468823] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.469047] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1026.469242] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1026.469439] env[63372]: INFO nova.compute.manager [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Took 1.51 seconds to destroy the instance on the hypervisor. [ 1026.469745] env[63372]: DEBUG oslo.service.loopingcall [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.470421] env[63372]: DEBUG nova.compute.manager [-] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1026.470521] env[63372]: DEBUG nova.network.neutron [-] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1026.477524] env[63372]: DEBUG nova.virt.hardware [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1026.477808] env[63372]: DEBUG nova.virt.hardware [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1026.477970] env[63372]: DEBUG nova.virt.hardware [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1026.478199] env[63372]: DEBUG nova.virt.hardware [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1026.478308] env[63372]: DEBUG nova.virt.hardware [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1026.478451] env[63372]: DEBUG nova.virt.hardware [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1026.478656] env[63372]: DEBUG nova.virt.hardware [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1026.478809] env[63372]: DEBUG nova.virt.hardware [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1026.479023] env[63372]: DEBUG nova.virt.hardware [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1026.479647] env[63372]: DEBUG nova.virt.hardware [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1026.479647] env[63372]: DEBUG nova.virt.hardware [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1026.480901] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-354d72df-e8b7-4522-944a-6005268ca94c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.491630] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28aadd98-38cf-4153-94c3-7004629cc91e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.506105] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.506393] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.507885] env[63372]: INFO nova.compute.claims [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1026.618274] env[63372]: DEBUG oslo_vmware.api [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024698, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133755} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.618274] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.618441] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1026.618496] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1026.638210] env[63372]: INFO nova.scheduler.client.report [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Deleted allocations for instance 91985614-b959-401e-bb06-d67b230ee026 [ 1026.814760] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024699, 'name': ReconfigVM_Task, 'duration_secs': 0.307751} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.817427] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Reconfigured VM instance instance-0000005a to attach disk [datastore2] b7732621-7e58-40a3-b723-5c66df6f74a1/b7732621-7e58-40a3-b723-5c66df6f74a1.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.817734] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance 'b7732621-7e58-40a3-b723-5c66df6f74a1' progress to 50 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1026.895175] env[63372]: DEBUG nova.objects.instance [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'pci_requests' on Instance uuid d7b7ee35-5e83-4c62-bd1e-8ec39951c44c {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.949322] env[63372]: DEBUG oslo_concurrency.lockutils [req-9b0a6a85-f3ce-49ca-b9fa-9291e165b974 req-37cc19ba-ad8c-450c-8d29-c2474d3445f5 service nova] Releasing lock "refresh_cache-b455b16b-3332-43bf-a91e-a350287ba5f3" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.142510] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.235951] env[63372]: DEBUG nova.network.neutron [-] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.325269] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e750cda7-53fd-441a-ae8c-fb52ffa0dab0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.352274] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20e6935-253d-41f2-931e-0630665e34f5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.375667] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance 'b7732621-7e58-40a3-b723-5c66df6f74a1' progress to 67 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1027.397398] env[63372]: DEBUG nova.objects.base [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1027.399168] env[63372]: DEBUG nova.network.neutron [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1027.468889] env[63372]: DEBUG nova.policy [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7ffe6776b664c00b9086046f9171ec6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46d2cb188a004dc385f2aa15458be929', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1027.503530] env[63372]: DEBUG nova.network.neutron [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Successfully updated port: 9281c23e-1f88-4862-8f9a-e3e3478592dc {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1027.564130] env[63372]: DEBUG nova.compute.manager [req-688f31a3-ace5-449d-8851-b86a49ede5b0 req-2ad86b68-9d90-4a93-a146-51ab19c64969 service nova] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Received event network-vif-plugged-9281c23e-1f88-4862-8f9a-e3e3478592dc {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1027.564130] env[63372]: DEBUG oslo_concurrency.lockutils [req-688f31a3-ace5-449d-8851-b86a49ede5b0 req-2ad86b68-9d90-4a93-a146-51ab19c64969 service nova] Acquiring lock "c38256cc-215c-4bec-94e5-24a1bbef3e10-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.564130] env[63372]: DEBUG oslo_concurrency.lockutils [req-688f31a3-ace5-449d-8851-b86a49ede5b0 req-2ad86b68-9d90-4a93-a146-51ab19c64969 service nova] Lock "c38256cc-215c-4bec-94e5-24a1bbef3e10-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.564130] env[63372]: DEBUG oslo_concurrency.lockutils [req-688f31a3-ace5-449d-8851-b86a49ede5b0 req-2ad86b68-9d90-4a93-a146-51ab19c64969 service nova] Lock "c38256cc-215c-4bec-94e5-24a1bbef3e10-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.564317] env[63372]: DEBUG nova.compute.manager [req-688f31a3-ace5-449d-8851-b86a49ede5b0 req-2ad86b68-9d90-4a93-a146-51ab19c64969 service nova] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] No waiting events found dispatching network-vif-plugged-9281c23e-1f88-4862-8f9a-e3e3478592dc {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1027.564493] env[63372]: WARNING nova.compute.manager [req-688f31a3-ace5-449d-8851-b86a49ede5b0 req-2ad86b68-9d90-4a93-a146-51ab19c64969 service nova] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Received unexpected event network-vif-plugged-9281c23e-1f88-4862-8f9a-e3e3478592dc for instance with vm_state building and task_state spawning. [ 1027.707030] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f3cf29-ad7d-44eb-921f-fa9c924f0c4c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.714829] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7538e8a-f0d7-4835-b967-3f85e7986932 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.465524] env[63372]: INFO nova.compute.manager [-] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Took 1.99 seconds to deallocate network for instance. [ 1028.470730] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "refresh_cache-c38256cc-215c-4bec-94e5-24a1bbef3e10" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.470865] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "refresh_cache-c38256cc-215c-4bec-94e5-24a1bbef3e10" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.471046] env[63372]: DEBUG nova.network.neutron [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1028.477417] env[63372]: DEBUG nova.compute.manager [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Received event network-changed-301c1926-329c-454a-9755-e33bef4550d8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.477598] env[63372]: DEBUG nova.compute.manager [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Refreshing instance network info cache due to event network-changed-301c1926-329c-454a-9755-e33bef4550d8. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1028.477797] env[63372]: DEBUG oslo_concurrency.lockutils [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] Acquiring lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.477931] env[63372]: DEBUG oslo_concurrency.lockutils [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] Acquired lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.478099] env[63372]: DEBUG nova.network.neutron [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Refreshing network info cache for port 301c1926-329c-454a-9755-e33bef4550d8 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1028.479476] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29fc973e-84fd-4a64-9c4f-2b7f01faf096 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.499753] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbde4edb-4ec0-4120-9d56-b484e87c71fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.511747] env[63372]: DEBUG nova.compute.provider_tree [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.850773] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.851066] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.985950] env[63372]: DEBUG oslo_concurrency.lockutils [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.986412] env[63372]: DEBUG oslo_concurrency.lockutils [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.986751] env[63372]: DEBUG oslo_concurrency.lockutils [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.987089] env[63372]: DEBUG oslo_concurrency.lockutils [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.987395] env[63372]: DEBUG oslo_concurrency.lockutils [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.990413] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.991054] env[63372]: INFO nova.compute.manager [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Terminating instance [ 1028.995853] env[63372]: DEBUG nova.compute.manager [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1028.996129] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1028.997250] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efbcb319-d684-4951-bd03-229b12c14885 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.003844] env[63372]: DEBUG nova.network.neutron [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1029.007737] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1029.007970] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a3c330f7-e40d-4eb9-ab5b-11004f2f7383 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.014470] env[63372]: DEBUG nova.scheduler.client.report [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.093106] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1029.093373] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1029.093667] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleting the datastore file [datastore1] 800cb48b-e75e-45e7-94dc-6ad147e6c6c5 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1029.094115] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7aa50ff6-f976-4526-a18c-8e091b3c0930 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.101194] env[63372]: DEBUG oslo_vmware.api [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 1029.101194] env[63372]: value = "task-1024701" [ 1029.101194] env[63372]: _type = "Task" [ 1029.101194] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.114829] env[63372]: DEBUG oslo_vmware.api [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024701, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.147136] env[63372]: DEBUG nova.network.neutron [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Updating instance_info_cache with network_info: [{"id": "9281c23e-1f88-4862-8f9a-e3e3478592dc", "address": "fa:16:3e:3e:40:1b", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9281c23e-1f", "ovs_interfaceid": "9281c23e-1f88-4862-8f9a-e3e3478592dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.220589] env[63372]: DEBUG nova.network.neutron [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Updated VIF entry in instance network info cache for port 301c1926-329c-454a-9755-e33bef4550d8. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1029.220965] env[63372]: DEBUG nova.network.neutron [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Updating instance_info_cache with network_info: [{"id": "301c1926-329c-454a-9755-e33bef4550d8", "address": "fa:16:3e:bc:d9:91", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap301c1926-32", "ovs_interfaceid": "301c1926-329c-454a-9755-e33bef4550d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.353659] env[63372]: DEBUG nova.compute.manager [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1029.518984] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.012s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.519516] env[63372]: DEBUG nova.compute.manager [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1029.522054] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.380s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.522312] env[63372]: DEBUG nova.objects.instance [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lazy-loading 'resources' on Instance uuid 91985614-b959-401e-bb06-d67b230ee026 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.588273] env[63372]: DEBUG nova.compute.manager [req-15fbcbb9-74ac-4c42-94a7-4facc074e849 req-ee9bd748-58a6-441c-910e-88dc0048fcdc service nova] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Received event network-changed-9281c23e-1f88-4862-8f9a-e3e3478592dc {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.588466] env[63372]: DEBUG nova.compute.manager [req-15fbcbb9-74ac-4c42-94a7-4facc074e849 req-ee9bd748-58a6-441c-910e-88dc0048fcdc service nova] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Refreshing instance network info cache due to event network-changed-9281c23e-1f88-4862-8f9a-e3e3478592dc. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1029.588660] env[63372]: DEBUG oslo_concurrency.lockutils [req-15fbcbb9-74ac-4c42-94a7-4facc074e849 req-ee9bd748-58a6-441c-910e-88dc0048fcdc service nova] Acquiring lock "refresh_cache-c38256cc-215c-4bec-94e5-24a1bbef3e10" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.613310] env[63372]: DEBUG oslo_vmware.api [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024701, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133698} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.613551] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.613760] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1029.613950] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1029.614144] env[63372]: INFO nova.compute.manager [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1029.614387] env[63372]: DEBUG oslo.service.loopingcall [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.614580] env[63372]: DEBUG nova.compute.manager [-] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1029.614671] env[63372]: DEBUG nova.network.neutron [-] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1029.620024] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "91985614-b959-401e-bb06-d67b230ee026" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.631776] env[63372]: DEBUG nova.network.neutron [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Port 128a0686-7ca8-427f-9cf7-c63a2f370e1f binding to destination host cpu-1 is already ACTIVE {{(pid=63372) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1029.649331] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "refresh_cache-c38256cc-215c-4bec-94e5-24a1bbef3e10" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.649619] env[63372]: DEBUG nova.compute.manager [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Instance network_info: |[{"id": "9281c23e-1f88-4862-8f9a-e3e3478592dc", "address": "fa:16:3e:3e:40:1b", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9281c23e-1f", "ovs_interfaceid": "9281c23e-1f88-4862-8f9a-e3e3478592dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1029.649915] env[63372]: DEBUG oslo_concurrency.lockutils [req-15fbcbb9-74ac-4c42-94a7-4facc074e849 req-ee9bd748-58a6-441c-910e-88dc0048fcdc service nova] Acquired lock "refresh_cache-c38256cc-215c-4bec-94e5-24a1bbef3e10" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.650098] env[63372]: DEBUG nova.network.neutron [req-15fbcbb9-74ac-4c42-94a7-4facc074e849 req-ee9bd748-58a6-441c-910e-88dc0048fcdc service nova] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Refreshing network info cache for port 9281c23e-1f88-4862-8f9a-e3e3478592dc {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1029.651287] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3e:40:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b21ab10-d886-4453-9472-9e11fb3c450d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9281c23e-1f88-4862-8f9a-e3e3478592dc', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1029.659247] env[63372]: DEBUG oslo.service.loopingcall [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.660197] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1029.660431] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-17ab2110-a5a3-4efe-b930-4b14a0732b65 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.680411] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1029.680411] env[63372]: value = "task-1024702" [ 1029.680411] env[63372]: _type = "Task" [ 1029.680411] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.688552] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024702, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.724010] env[63372]: DEBUG oslo_concurrency.lockutils [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] Releasing lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.724311] env[63372]: DEBUG nova.compute.manager [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Received event network-vif-unplugged-57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.724514] env[63372]: DEBUG oslo_concurrency.lockutils [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] Acquiring lock "91985614-b959-401e-bb06-d67b230ee026-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.724749] env[63372]: DEBUG oslo_concurrency.lockutils [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] Lock "91985614-b959-401e-bb06-d67b230ee026-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.724937] env[63372]: DEBUG oslo_concurrency.lockutils [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] Lock "91985614-b959-401e-bb06-d67b230ee026-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.725125] env[63372]: DEBUG nova.compute.manager [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] No waiting events found dispatching network-vif-unplugged-57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1029.725302] env[63372]: WARNING nova.compute.manager [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Received unexpected event network-vif-unplugged-57943c40-a8c1-4d1a-bb83-d257029ce77f for instance with vm_state shelved_offloaded and task_state None. [ 1029.725466] env[63372]: DEBUG nova.compute.manager [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Received event network-changed-57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.725698] env[63372]: DEBUG nova.compute.manager [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Refreshing instance network info cache due to event network-changed-57943c40-a8c1-4d1a-bb83-d257029ce77f. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1029.725942] env[63372]: DEBUG oslo_concurrency.lockutils [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] Acquiring lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.726242] env[63372]: DEBUG oslo_concurrency.lockutils [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] Acquired lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.726447] env[63372]: DEBUG nova.network.neutron [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Refreshing network info cache for port 57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1029.876684] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.999237] env[63372]: DEBUG nova.network.neutron [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Successfully updated port: 1d17bf13-a03b-4654-bdfe-adc2d47e273a {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1030.025934] env[63372]: DEBUG nova.compute.utils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1030.028708] env[63372]: DEBUG nova.objects.instance [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lazy-loading 'numa_topology' on Instance uuid 91985614-b959-401e-bb06-d67b230ee026 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.029983] env[63372]: DEBUG nova.compute.manager [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1030.030172] env[63372]: DEBUG nova.network.neutron [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1030.069103] env[63372]: DEBUG nova.policy [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2253a9a409d4f1ba9176485655c0988', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dac773c2536745e285181426ae34bb96', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1030.194948] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024702, 'name': CreateVM_Task, 'duration_secs': 0.338393} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.195141] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1030.195790] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.195951] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.196288] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1030.196587] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03787a1b-fa0c-4dbc-9be0-0acd6efbdba3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.201156] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1030.201156] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5283aeac-f593-d993-1424-4a18da579b49" [ 1030.201156] env[63372]: _type = "Task" [ 1030.201156] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.211099] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5283aeac-f593-d993-1424-4a18da579b49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.370959] env[63372]: DEBUG nova.network.neutron [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Successfully created port: ebc3af7a-29a8-44e6-85e7-4ae1bafa1844 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1030.395750] env[63372]: DEBUG nova.network.neutron [req-15fbcbb9-74ac-4c42-94a7-4facc074e849 req-ee9bd748-58a6-441c-910e-88dc0048fcdc service nova] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Updated VIF entry in instance network info cache for port 9281c23e-1f88-4862-8f9a-e3e3478592dc. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1030.396161] env[63372]: DEBUG nova.network.neutron [req-15fbcbb9-74ac-4c42-94a7-4facc074e849 req-ee9bd748-58a6-441c-910e-88dc0048fcdc service nova] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Updating instance_info_cache with network_info: [{"id": "9281c23e-1f88-4862-8f9a-e3e3478592dc", "address": "fa:16:3e:3e:40:1b", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9281c23e-1f", "ovs_interfaceid": "9281c23e-1f88-4862-8f9a-e3e3478592dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.427134] env[63372]: DEBUG nova.compute.manager [req-9936c926-1140-4bff-949e-940e05e3df79 req-eaa8cbea-8ddc-432d-9c6a-7621eeee2088 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Received event network-vif-plugged-1d17bf13-a03b-4654-bdfe-adc2d47e273a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.427372] env[63372]: DEBUG oslo_concurrency.lockutils [req-9936c926-1140-4bff-949e-940e05e3df79 req-eaa8cbea-8ddc-432d-9c6a-7621eeee2088 service nova] Acquiring lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.427585] env[63372]: DEBUG oslo_concurrency.lockutils [req-9936c926-1140-4bff-949e-940e05e3df79 req-eaa8cbea-8ddc-432d-9c6a-7621eeee2088 service nova] Lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.427755] env[63372]: DEBUG oslo_concurrency.lockutils [req-9936c926-1140-4bff-949e-940e05e3df79 req-eaa8cbea-8ddc-432d-9c6a-7621eeee2088 service nova] Lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.427926] env[63372]: DEBUG nova.compute.manager [req-9936c926-1140-4bff-949e-940e05e3df79 req-eaa8cbea-8ddc-432d-9c6a-7621eeee2088 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] No waiting events found dispatching network-vif-plugged-1d17bf13-a03b-4654-bdfe-adc2d47e273a {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1030.428196] env[63372]: WARNING nova.compute.manager [req-9936c926-1140-4bff-949e-940e05e3df79 req-eaa8cbea-8ddc-432d-9c6a-7621eeee2088 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Received unexpected event network-vif-plugged-1d17bf13-a03b-4654-bdfe-adc2d47e273a for instance with vm_state active and task_state None. [ 1030.428394] env[63372]: DEBUG nova.compute.manager [req-9936c926-1140-4bff-949e-940e05e3df79 req-eaa8cbea-8ddc-432d-9c6a-7621eeee2088 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Received event network-changed-1d17bf13-a03b-4654-bdfe-adc2d47e273a {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.428556] env[63372]: DEBUG nova.compute.manager [req-9936c926-1140-4bff-949e-940e05e3df79 req-eaa8cbea-8ddc-432d-9c6a-7621eeee2088 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Refreshing instance network info cache due to event network-changed-1d17bf13-a03b-4654-bdfe-adc2d47e273a. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1030.428742] env[63372]: DEBUG oslo_concurrency.lockutils [req-9936c926-1140-4bff-949e-940e05e3df79 req-eaa8cbea-8ddc-432d-9c6a-7621eeee2088 service nova] Acquiring lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.428881] env[63372]: DEBUG oslo_concurrency.lockutils [req-9936c926-1140-4bff-949e-940e05e3df79 req-eaa8cbea-8ddc-432d-9c6a-7621eeee2088 service nova] Acquired lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.429069] env[63372]: DEBUG nova.network.neutron [req-9936c926-1140-4bff-949e-940e05e3df79 req-eaa8cbea-8ddc-432d-9c6a-7621eeee2088 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Refreshing network info cache for port 1d17bf13-a03b-4654-bdfe-adc2d47e273a {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1030.436989] env[63372]: DEBUG nova.network.neutron [-] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.501872] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.531586] env[63372]: DEBUG nova.compute.manager [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1030.532784] env[63372]: DEBUG nova.objects.base [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Object Instance<91985614-b959-401e-bb06-d67b230ee026> lazy-loaded attributes: resources,numa_topology {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1030.664152] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "b7732621-7e58-40a3-b723-5c66df6f74a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.664548] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "b7732621-7e58-40a3-b723-5c66df6f74a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.664853] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "b7732621-7e58-40a3-b723-5c66df6f74a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.717870] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5283aeac-f593-d993-1424-4a18da579b49, 'name': SearchDatastore_Task, 'duration_secs': 0.009718} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.720596] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.720889] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1030.721201] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.721399] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.721609] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1030.722103] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-751aa425-6cfc-4853-86d6-a627fabf97ea {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.735237] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1030.735404] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1030.736922] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aec44dbf-340b-49eb-a914-bb5116cb44d6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.743114] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1030.743114] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fd138a-0cb5-705a-3191-f352ad108a06" [ 1030.743114] env[63372]: _type = "Task" [ 1030.743114] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.753898] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fd138a-0cb5-705a-3191-f352ad108a06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.770448] env[63372]: DEBUG nova.network.neutron [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updated VIF entry in instance network info cache for port 57943c40-a8c1-4d1a-bb83-d257029ce77f. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1030.771348] env[63372]: DEBUG nova.network.neutron [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updating instance_info_cache with network_info: [{"id": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "address": "fa:16:3e:1f:74:b1", "network": {"id": "cbdcb945-f0a5-49b1-9b18-fcd8e4ca9728", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1068757189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e517915d5e7a4e3e924cc95c55170a66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap57943c40-a8", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.805968] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9669d222-6f60-4ed6-8788-f1ec4c4f578f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.816121] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c73c3e9-f4a4-4110-bb9e-c0693309a24f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.847147] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b336db-d9f2-4dc8-9fc4-657bd853ae6c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.855280] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e393e6-4ba1-48be-8195-7f10240446fe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.868573] env[63372]: DEBUG nova.compute.provider_tree [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.899288] env[63372]: DEBUG oslo_concurrency.lockutils [req-15fbcbb9-74ac-4c42-94a7-4facc074e849 req-ee9bd748-58a6-441c-910e-88dc0048fcdc service nova] Releasing lock "refresh_cache-c38256cc-215c-4bec-94e5-24a1bbef3e10" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.939486] env[63372]: INFO nova.compute.manager [-] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Took 1.32 seconds to deallocate network for instance. [ 1031.134298] env[63372]: DEBUG nova.network.neutron [req-9936c926-1140-4bff-949e-940e05e3df79 req-eaa8cbea-8ddc-432d-9c6a-7621eeee2088 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Added VIF to instance network info cache for port 1d17bf13-a03b-4654-bdfe-adc2d47e273a. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3489}} [ 1031.134744] env[63372]: DEBUG nova.network.neutron [req-9936c926-1140-4bff-949e-940e05e3df79 req-eaa8cbea-8ddc-432d-9c6a-7621eeee2088 service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Updating instance_info_cache with network_info: [{"id": "301c1926-329c-454a-9755-e33bef4550d8", "address": "fa:16:3e:bc:d9:91", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap301c1926-32", "ovs_interfaceid": "301c1926-329c-454a-9755-e33bef4550d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1d17bf13-a03b-4654-bdfe-adc2d47e273a", "address": "fa:16:3e:09:36:65", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d17bf13-a0", "ovs_interfaceid": "1d17bf13-a03b-4654-bdfe-adc2d47e273a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.254544] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fd138a-0cb5-705a-3191-f352ad108a06, 'name': SearchDatastore_Task, 'duration_secs': 0.011298} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.255174] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6673fdcd-10e6-45c3-8a2c-62c5eb864080 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.261113] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1031.261113] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52580961-ee38-1318-4ea0-1feb134e1eb0" [ 1031.261113] env[63372]: _type = "Task" [ 1031.261113] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.268860] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52580961-ee38-1318-4ea0-1feb134e1eb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.273475] env[63372]: DEBUG oslo_concurrency.lockutils [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] Releasing lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.273727] env[63372]: DEBUG nova.compute.manager [req-4494612e-c156-46ab-8f7d-ba0413d0f285 req-f10eec47-bb1e-4fd7-90e8-bea98b8548cb service nova] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Received event network-vif-deleted-298d217f-c17a-4ee2-acae-baee355112dd {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.372123] env[63372]: DEBUG nova.scheduler.client.report [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1031.445946] env[63372]: DEBUG oslo_concurrency.lockutils [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.538652] env[63372]: DEBUG nova.compute.manager [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1031.565338] env[63372]: DEBUG nova.virt.hardware [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.565595] env[63372]: DEBUG nova.virt.hardware [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.565758] env[63372]: DEBUG nova.virt.hardware [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.565936] env[63372]: DEBUG nova.virt.hardware [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.566119] env[63372]: DEBUG nova.virt.hardware [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.566290] env[63372]: DEBUG nova.virt.hardware [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.566496] env[63372]: DEBUG nova.virt.hardware [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.566651] env[63372]: DEBUG nova.virt.hardware [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.566812] env[63372]: DEBUG nova.virt.hardware [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.566967] env[63372]: DEBUG nova.virt.hardware [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.567146] env[63372]: DEBUG nova.virt.hardware [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.567989] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b011df-418a-4b03-85cf-e08633cc78ef {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.575676] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e274d3-379a-4a34-9f0c-dd0ca7040177 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.619836] env[63372]: DEBUG nova.compute.manager [req-dac309c9-a72f-4862-afb8-529adf38c5de req-a482243e-a015-4d51-bb2c-b09924f5b178 service nova] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Received event network-vif-deleted-f96d5780-3a85-4214-8d8e-2b02ca1c4c2e {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.638929] env[63372]: DEBUG oslo_concurrency.lockutils [req-9936c926-1140-4bff-949e-940e05e3df79 req-eaa8cbea-8ddc-432d-9c6a-7621eeee2088 service nova] Releasing lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.639797] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.639797] env[63372]: DEBUG nova.network.neutron [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1031.700885] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.701124] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.701309] env[63372]: DEBUG nova.network.neutron [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1031.770931] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52580961-ee38-1318-4ea0-1feb134e1eb0, 'name': SearchDatastore_Task, 'duration_secs': 0.013028} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.771238] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.771492] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] c38256cc-215c-4bec-94e5-24a1bbef3e10/c38256cc-215c-4bec-94e5-24a1bbef3e10.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1031.771732] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d7cd81ad-e6f5-4365-8caa-554b3297cdf5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.778951] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1031.778951] env[63372]: value = "task-1024703" [ 1031.778951] env[63372]: _type = "Task" [ 1031.778951] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.786339] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024703, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.878194] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.356s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.880740] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.890s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.880948] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.883078] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.006s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.884509] env[63372]: INFO nova.compute.claims [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1031.908808] env[63372]: INFO nova.scheduler.client.report [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleted allocations for instance cbb471c5-8d86-4d8f-8935-f6a77e5144be [ 1032.194326] env[63372]: WARNING nova.network.neutron [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] e1f7ca1b-541e-468f-a439-a6841fc781ba already exists in list: networks containing: ['e1f7ca1b-541e-468f-a439-a6841fc781ba']. ignoring it [ 1032.194583] env[63372]: WARNING nova.network.neutron [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] e1f7ca1b-541e-468f-a439-a6841fc781ba already exists in list: networks containing: ['e1f7ca1b-541e-468f-a439-a6841fc781ba']. ignoring it [ 1032.194788] env[63372]: WARNING nova.network.neutron [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] 1d17bf13-a03b-4654-bdfe-adc2d47e273a already exists in list: port_ids containing: ['1d17bf13-a03b-4654-bdfe-adc2d47e273a']. ignoring it [ 1032.209409] env[63372]: DEBUG nova.network.neutron [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Successfully updated port: ebc3af7a-29a8-44e6-85e7-4ae1bafa1844 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1032.289805] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024703, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485383} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.292861] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] c38256cc-215c-4bec-94e5-24a1bbef3e10/c38256cc-215c-4bec-94e5-24a1bbef3e10.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1032.293116] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1032.293672] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dcf2f8c0-8729-4e35-bdc3-b5bb0ecc7fcd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.300838] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1032.300838] env[63372]: value = "task-1024704" [ 1032.300838] env[63372]: _type = "Task" [ 1032.300838] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.308837] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024704, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.395504] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c2d0b7e8-ea14-4447-9c4f-da24efad0bbe tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "91985614-b959-401e-bb06-d67b230ee026" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 30.661s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.396450] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "91985614-b959-401e-bb06-d67b230ee026" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.776s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.396648] env[63372]: INFO nova.compute.manager [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Unshelving [ 1032.418358] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d5a81f5e-4cb9-493b-a713-2c075cefd2ba tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "cbb471c5-8d86-4d8f-8935-f6a77e5144be" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.468s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.516730] env[63372]: DEBUG nova.network.neutron [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Updating instance_info_cache with network_info: [{"id": "301c1926-329c-454a-9755-e33bef4550d8", "address": "fa:16:3e:bc:d9:91", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap301c1926-32", "ovs_interfaceid": "301c1926-329c-454a-9755-e33bef4550d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1d17bf13-a03b-4654-bdfe-adc2d47e273a", "address": "fa:16:3e:09:36:65", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d17bf13-a0", "ovs_interfaceid": "1d17bf13-a03b-4654-bdfe-adc2d47e273a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.544314] env[63372]: DEBUG nova.network.neutron [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance_info_cache with network_info: [{"id": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "address": "fa:16:3e:28:02:b8", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap128a0686-7c", "ovs_interfaceid": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.609226] env[63372]: DEBUG nova.compute.manager [req-a239c86f-7c29-49f6-af6b-05eb0571e87e req-23132cff-8fd5-45d8-9171-9855468d8a1c service nova] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Received event network-vif-plugged-ebc3af7a-29a8-44e6-85e7-4ae1bafa1844 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.609456] env[63372]: DEBUG oslo_concurrency.lockutils [req-a239c86f-7c29-49f6-af6b-05eb0571e87e req-23132cff-8fd5-45d8-9171-9855468d8a1c service nova] Acquiring lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.609696] env[63372]: DEBUG oslo_concurrency.lockutils [req-a239c86f-7c29-49f6-af6b-05eb0571e87e req-23132cff-8fd5-45d8-9171-9855468d8a1c service nova] Lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.609824] env[63372]: DEBUG oslo_concurrency.lockutils [req-a239c86f-7c29-49f6-af6b-05eb0571e87e req-23132cff-8fd5-45d8-9171-9855468d8a1c service nova] Lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.609993] env[63372]: DEBUG nova.compute.manager [req-a239c86f-7c29-49f6-af6b-05eb0571e87e req-23132cff-8fd5-45d8-9171-9855468d8a1c service nova] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] No waiting events found dispatching network-vif-plugged-ebc3af7a-29a8-44e6-85e7-4ae1bafa1844 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1032.610178] env[63372]: WARNING nova.compute.manager [req-a239c86f-7c29-49f6-af6b-05eb0571e87e req-23132cff-8fd5-45d8-9171-9855468d8a1c service nova] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Received unexpected event network-vif-plugged-ebc3af7a-29a8-44e6-85e7-4ae1bafa1844 for instance with vm_state building and task_state spawning. [ 1032.610352] env[63372]: DEBUG nova.compute.manager [req-a239c86f-7c29-49f6-af6b-05eb0571e87e req-23132cff-8fd5-45d8-9171-9855468d8a1c service nova] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Received event network-changed-ebc3af7a-29a8-44e6-85e7-4ae1bafa1844 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.610512] env[63372]: DEBUG nova.compute.manager [req-a239c86f-7c29-49f6-af6b-05eb0571e87e req-23132cff-8fd5-45d8-9171-9855468d8a1c service nova] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Refreshing instance network info cache due to event network-changed-ebc3af7a-29a8-44e6-85e7-4ae1bafa1844. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1032.610692] env[63372]: DEBUG oslo_concurrency.lockutils [req-a239c86f-7c29-49f6-af6b-05eb0571e87e req-23132cff-8fd5-45d8-9171-9855468d8a1c service nova] Acquiring lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.610831] env[63372]: DEBUG oslo_concurrency.lockutils [req-a239c86f-7c29-49f6-af6b-05eb0571e87e req-23132cff-8fd5-45d8-9171-9855468d8a1c service nova] Acquired lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.610987] env[63372]: DEBUG nova.network.neutron [req-a239c86f-7c29-49f6-af6b-05eb0571e87e req-23132cff-8fd5-45d8-9171-9855468d8a1c service nova] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Refreshing network info cache for port ebc3af7a-29a8-44e6-85e7-4ae1bafa1844 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1032.713494] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.810700] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024704, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068232} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.811573] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1032.811976] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f3c059-714e-4b59-a95a-f3f33b15ba19 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.833910] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] c38256cc-215c-4bec-94e5-24a1bbef3e10/c38256cc-215c-4bec-94e5-24a1bbef3e10.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.834209] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5b5061d-4c30-483a-ab46-500814303db5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.853815] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1032.853815] env[63372]: value = "task-1024705" [ 1032.853815] env[63372]: _type = "Task" [ 1032.853815] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.861740] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024705, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.019661] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.020371] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.020538] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.021615] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bc4a37c-cebc-47d7-b63a-4ba7a61c7477 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.038766] env[63372]: DEBUG nova.virt.hardware [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1033.039050] env[63372]: DEBUG nova.virt.hardware [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1033.039249] env[63372]: DEBUG nova.virt.hardware [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1033.039454] env[63372]: DEBUG nova.virt.hardware [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1033.039609] env[63372]: DEBUG nova.virt.hardware [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1033.039776] env[63372]: DEBUG nova.virt.hardware [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1033.039995] env[63372]: DEBUG nova.virt.hardware [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1033.040188] env[63372]: DEBUG nova.virt.hardware [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1033.040374] env[63372]: DEBUG nova.virt.hardware [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1033.040596] env[63372]: DEBUG nova.virt.hardware [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1033.040790] env[63372]: DEBUG nova.virt.hardware [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1033.047080] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Reconfiguring VM to attach interface {{(pid=63372) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1033.049878] env[63372]: DEBUG oslo_concurrency.lockutils [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.053323] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f68f862-6ca5-4652-a65c-618bf7dae0c7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.071791] env[63372]: DEBUG oslo_vmware.api [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 1033.071791] env[63372]: value = "task-1024706" [ 1033.071791] env[63372]: _type = "Task" [ 1033.071791] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.082223] env[63372]: DEBUG oslo_vmware.api [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024706, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.111371] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db764575-f659-4b4c-9d8b-2beff387e8e0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.121570] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb8e948-edae-4c45-a4cf-d009fadc0c55 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.154742] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b3ca3f-9680-4d55-a168-e0ae10c82bbc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.158185] env[63372]: DEBUG nova.network.neutron [req-a239c86f-7c29-49f6-af6b-05eb0571e87e req-23132cff-8fd5-45d8-9171-9855468d8a1c service nova] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1033.165374] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4402d715-1884-4f0f-b97a-0a9382dc5451 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.180228] env[63372]: DEBUG nova.compute.provider_tree [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.266534] env[63372]: DEBUG nova.network.neutron [req-a239c86f-7c29-49f6-af6b-05eb0571e87e req-23132cff-8fd5-45d8-9171-9855468d8a1c service nova] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.364835] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.418930] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.570287] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03eb605a-9cf9-4981-846c-d0ef28c6850f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.581257] env[63372]: DEBUG oslo_vmware.api [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024706, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.583262] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038bc751-8003-4286-afd4-109133980950 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.683061] env[63372]: DEBUG nova.scheduler.client.report [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1033.769719] env[63372]: DEBUG oslo_concurrency.lockutils [req-a239c86f-7c29-49f6-af6b-05eb0571e87e req-23132cff-8fd5-45d8-9171-9855468d8a1c service nova] Releasing lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.770085] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.770255] env[63372]: DEBUG nova.network.neutron [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1033.865553] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024705, 'name': ReconfigVM_Task, 'duration_secs': 0.684508} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.865801] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Reconfigured VM instance instance-00000066 to attach disk [datastore2] c38256cc-215c-4bec-94e5-24a1bbef3e10/c38256cc-215c-4bec-94e5-24a1bbef3e10.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.866449] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2508fdf6-a8dd-4bcb-9e82-a98add54e8c2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.872616] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1033.872616] env[63372]: value = "task-1024707" [ 1033.872616] env[63372]: _type = "Task" [ 1033.872616] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.882236] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024707, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.082549] env[63372]: DEBUG oslo_vmware.api [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024706, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.188039] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.305s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.188567] env[63372]: DEBUG nova.compute.manager [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1034.191138] env[63372]: DEBUG oslo_concurrency.lockutils [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.745s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.191365] env[63372]: DEBUG nova.objects.instance [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lazy-loading 'resources' on Instance uuid 800cb48b-e75e-45e7-94dc-6ad147e6c6c5 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.315991] env[63372]: DEBUG nova.network.neutron [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1034.382610] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024707, 'name': Rename_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.502931] env[63372]: DEBUG nova.network.neutron [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating instance_info_cache with network_info: [{"id": "ebc3af7a-29a8-44e6-85e7-4ae1bafa1844", "address": "fa:16:3e:05:a3:85", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebc3af7a-29", "ovs_interfaceid": "ebc3af7a-29a8-44e6-85e7-4ae1bafa1844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.583197] env[63372]: DEBUG oslo_vmware.api [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024706, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.681916] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a6540d3-0a52-42aa-9289-34cbbc6dc5c3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.706292] env[63372]: DEBUG nova.compute.utils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1034.708593] env[63372]: DEBUG nova.compute.manager [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1034.708593] env[63372]: DEBUG nova.network.neutron [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1034.710552] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84da58c3-ac21-4fb1-af2c-a6a29de4bfc1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.718277] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance 'b7732621-7e58-40a3-b723-5c66df6f74a1' progress to 83 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1034.751748] env[63372]: DEBUG nova.policy [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc9dec0b991f4f61a674a2a844b5462c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ca1f09f7f9e4b2b8010e478202373ea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1034.886387] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024707, 'name': Rename_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.892594] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b808cc-86c8-4689-9adb-74a0bc6faa6a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.900242] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32344e91-cade-4038-bc38-6247e2976989 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.934581] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b520243-9a62-4a7b-9126-5f18afe8b7c0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.942550] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0086930f-b811-44cf-a505-e6ea1b11bffb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.956819] env[63372]: DEBUG nova.compute.provider_tree [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.005645] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.005767] env[63372]: DEBUG nova.compute.manager [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Instance network_info: |[{"id": "ebc3af7a-29a8-44e6-85e7-4ae1bafa1844", "address": "fa:16:3e:05:a3:85", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebc3af7a-29", "ovs_interfaceid": "ebc3af7a-29a8-44e6-85e7-4ae1bafa1844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1035.006270] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:a3:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d4ef133-b6f3-41d1-add4-92a1482195cf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ebc3af7a-29a8-44e6-85e7-4ae1bafa1844', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1035.015562] env[63372]: DEBUG oslo.service.loopingcall [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.015870] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1035.016047] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d8e8502-0154-4f3a-8bca-89e54e551943 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.037030] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1035.037030] env[63372]: value = "task-1024708" [ 1035.037030] env[63372]: _type = "Task" [ 1035.037030] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.045215] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024708, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.079153] env[63372]: DEBUG nova.network.neutron [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Successfully created port: 0fbd29da-197e-452a-9644-ab461d995ec3 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1035.086228] env[63372]: DEBUG oslo_vmware.api [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024706, 'name': ReconfigVM_Task, 'duration_secs': 1.896761} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.086733] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.086970] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Reconfigured VM to attach interface {{(pid=63372) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1035.214658] env[63372]: DEBUG nova.compute.manager [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1035.225094] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1035.225094] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7b27e48-368a-4633-9a8e-3687954d54f9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.231666] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1035.231666] env[63372]: value = "task-1024709" [ 1035.231666] env[63372]: _type = "Task" [ 1035.231666] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.240251] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024709, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.383924] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024707, 'name': Rename_Task, 'duration_secs': 1.117344} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.384227] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1035.384512] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6363ec34-0d9a-4515-8033-a447f3deb730 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.390610] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1035.390610] env[63372]: value = "task-1024710" [ 1035.390610] env[63372]: _type = "Task" [ 1035.390610] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.401142] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024710, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.460577] env[63372]: DEBUG nova.scheduler.client.report [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1035.550220] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024708, 'name': CreateVM_Task, 'duration_secs': 0.378522} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.550450] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1035.551321] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.551537] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.552060] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1035.552391] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-032e7615-b14c-4d5c-a753-4356bd254630 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.558758] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1035.558758] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52c5e14d-a094-76a0-f2c6-d4fe96c511c6" [ 1035.558758] env[63372]: _type = "Task" [ 1035.558758] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.567106] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c5e14d-a094-76a0-f2c6-d4fe96c511c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.592801] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fa9b78e6-abf4-4b68-928d-1996dd647f32 tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-1d17bf13-a03b-4654-bdfe-adc2d47e273a" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 9.323s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.741852] env[63372]: DEBUG oslo_vmware.api [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024709, 'name': PowerOnVM_Task, 'duration_secs': 0.412236} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.742196] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1035.742403] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-30880716-dd19-4244-938d-9a9eebd59350 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance 'b7732621-7e58-40a3-b723-5c66df6f74a1' progress to 100 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1035.901119] env[63372]: DEBUG oslo_vmware.api [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024710, 'name': PowerOnVM_Task, 'duration_secs': 0.469453} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.901441] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1035.901651] env[63372]: INFO nova.compute.manager [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Took 9.46 seconds to spawn the instance on the hypervisor. [ 1035.901834] env[63372]: DEBUG nova.compute.manager [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1035.902647] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c67c0bd-ae13-4110-b258-c26b198ffa40 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.965247] env[63372]: DEBUG oslo_concurrency.lockutils [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.774s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.967525] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.549s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.967753] env[63372]: DEBUG nova.objects.instance [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lazy-loading 'pci_requests' on Instance uuid 91985614-b959-401e-bb06-d67b230ee026 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.989066] env[63372]: INFO nova.scheduler.client.report [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleted allocations for instance 800cb48b-e75e-45e7-94dc-6ad147e6c6c5 [ 1036.068400] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c5e14d-a094-76a0-f2c6-d4fe96c511c6, 'name': SearchDatastore_Task, 'duration_secs': 0.010466} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.068727] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.068970] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1036.069249] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.069402] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.069585] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1036.070100] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6eee9aa1-7a50-4ac7-822b-3d25cc9e4fab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.083259] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1036.083440] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1036.084152] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f6278b0-f076-43dd-8c0e-1bc439d4934f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.089117] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1036.089117] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5237b332-6979-786f-063b-172f377e6a03" [ 1036.089117] env[63372]: _type = "Task" [ 1036.089117] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.097178] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5237b332-6979-786f-063b-172f377e6a03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.224873] env[63372]: DEBUG nova.compute.manager [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1036.253892] env[63372]: DEBUG nova.virt.hardware [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1036.254030] env[63372]: DEBUG nova.virt.hardware [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1036.254147] env[63372]: DEBUG nova.virt.hardware [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1036.254333] env[63372]: DEBUG nova.virt.hardware [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1036.254483] env[63372]: DEBUG nova.virt.hardware [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1036.254634] env[63372]: DEBUG nova.virt.hardware [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1036.254841] env[63372]: DEBUG nova.virt.hardware [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1036.255017] env[63372]: DEBUG nova.virt.hardware [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1036.255196] env[63372]: DEBUG nova.virt.hardware [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1036.255366] env[63372]: DEBUG nova.virt.hardware [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1036.255543] env[63372]: DEBUG nova.virt.hardware [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1036.256408] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2591f8c-cab0-49ef-94a4-13e8093d82c0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.264652] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e260e85f-92e2-4ca8-a986-3d14df317ca8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.428951] env[63372]: INFO nova.compute.manager [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Took 14.76 seconds to build instance. [ 1036.471221] env[63372]: DEBUG nova.objects.instance [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lazy-loading 'numa_topology' on Instance uuid 91985614-b959-401e-bb06-d67b230ee026 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.497392] env[63372]: DEBUG oslo_concurrency.lockutils [None req-acd59fe2-593d-44ec-847f-5b14f39af64e tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "800cb48b-e75e-45e7-94dc-6ad147e6c6c5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.511s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.510823] env[63372]: DEBUG nova.compute.manager [req-545825db-8b87-4ec2-b970-fd62fe190bc0 req-fd3d9342-9291-4721-b016-244cbbd22ad5 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Received event network-vif-plugged-0fbd29da-197e-452a-9644-ab461d995ec3 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1036.511399] env[63372]: DEBUG oslo_concurrency.lockutils [req-545825db-8b87-4ec2-b970-fd62fe190bc0 req-fd3d9342-9291-4721-b016-244cbbd22ad5 service nova] Acquiring lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.511681] env[63372]: DEBUG oslo_concurrency.lockutils [req-545825db-8b87-4ec2-b970-fd62fe190bc0 req-fd3d9342-9291-4721-b016-244cbbd22ad5 service nova] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.511864] env[63372]: DEBUG oslo_concurrency.lockutils [req-545825db-8b87-4ec2-b970-fd62fe190bc0 req-fd3d9342-9291-4721-b016-244cbbd22ad5 service nova] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.512058] env[63372]: DEBUG nova.compute.manager [req-545825db-8b87-4ec2-b970-fd62fe190bc0 req-fd3d9342-9291-4721-b016-244cbbd22ad5 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] No waiting events found dispatching network-vif-plugged-0fbd29da-197e-452a-9644-ab461d995ec3 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1036.512463] env[63372]: WARNING nova.compute.manager [req-545825db-8b87-4ec2-b970-fd62fe190bc0 req-fd3d9342-9291-4721-b016-244cbbd22ad5 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Received unexpected event network-vif-plugged-0fbd29da-197e-452a-9644-ab461d995ec3 for instance with vm_state building and task_state spawning. [ 1036.598992] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5237b332-6979-786f-063b-172f377e6a03, 'name': SearchDatastore_Task, 'duration_secs': 0.022236} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.599846] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd45edca-f5ba-4f45-91da-e0368f491b4a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.604813] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1036.604813] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520aab90-6fa4-2c8e-1040-017394438f1c" [ 1036.604813] env[63372]: _type = "Task" [ 1036.604813] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.611987] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520aab90-6fa4-2c8e-1040-017394438f1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.655402] env[63372]: DEBUG nova.network.neutron [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Successfully updated port: 0fbd29da-197e-452a-9644-ab461d995ec3 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1036.931776] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0e3e4dca-13ec-4593-af9f-454fcc601d8e tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c38256cc-215c-4bec-94e5-24a1bbef3e10" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.268s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.975302] env[63372]: INFO nova.compute.claims [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1037.116376] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520aab90-6fa4-2c8e-1040-017394438f1c, 'name': SearchDatastore_Task, 'duration_secs': 0.025022} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.116376] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.116376] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d/83db51f8-2c4b-4ef1-afcb-ecac63bcab9d.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1037.116376] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fda83955-7f9e-4d82-9824-97390cccc799 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.122435] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1037.122435] env[63372]: value = "task-1024711" [ 1037.122435] env[63372]: _type = "Task" [ 1037.122435] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.130099] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024711, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.158439] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.158603] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.158760] env[63372]: DEBUG nova.network.neutron [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1037.362591] env[63372]: DEBUG oslo_concurrency.lockutils [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "interface-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-1d17bf13-a03b-4654-bdfe-adc2d47e273a" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.362870] env[63372]: DEBUG oslo_concurrency.lockutils [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-1d17bf13-a03b-4654-bdfe-adc2d47e273a" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.366780] env[63372]: DEBUG oslo_concurrency.lockutils [None req-edafe718-5fac-49ee-a4de-fd758143f31c tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "c38256cc-215c-4bec-94e5-24a1bbef3e10" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.367014] env[63372]: DEBUG oslo_concurrency.lockutils [None req-edafe718-5fac-49ee-a4de-fd758143f31c tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c38256cc-215c-4bec-94e5-24a1bbef3e10" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.367202] env[63372]: DEBUG nova.compute.manager [None req-edafe718-5fac-49ee-a4de-fd758143f31c tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1037.368083] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1490888-6f15-4a83-be83-a719596c03ca {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.376275] env[63372]: DEBUG nova.compute.manager [None req-edafe718-5fac-49ee-a4de-fd758143f31c tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63372) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1037.376875] env[63372]: DEBUG nova.objects.instance [None req-edafe718-5fac-49ee-a4de-fd758143f31c tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lazy-loading 'flavor' on Instance uuid c38256cc-215c-4bec-94e5-24a1bbef3e10 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.631851] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024711, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.689082] env[63372]: DEBUG nova.network.neutron [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1037.822022] env[63372]: DEBUG nova.network.neutron [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Updating instance_info_cache with network_info: [{"id": "0fbd29da-197e-452a-9644-ab461d995ec3", "address": "fa:16:3e:c8:e0:b1", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fbd29da-19", "ovs_interfaceid": "0fbd29da-197e-452a-9644-ab461d995ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.866526] env[63372]: DEBUG oslo_concurrency.lockutils [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.866716] env[63372]: DEBUG oslo_concurrency.lockutils [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.867926] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6edc390e-a932-4825-8d24-bf5699fcc5c9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.888515] env[63372]: DEBUG nova.network.neutron [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Port 128a0686-7ca8-427f-9cf7-c63a2f370e1f binding to destination host cpu-1 is already ACTIVE {{(pid=63372) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1037.888780] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.888931] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.890087] env[63372]: DEBUG nova.network.neutron [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1037.890927] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a3d901-30d9-4618-9260-557cea500ab5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.893734] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "c4718797-aa86-4ec0-94d3-6480bd6aa898" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.893972] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "c4718797-aa86-4ec0-94d3-6480bd6aa898" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.894194] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "c4718797-aa86-4ec0-94d3-6480bd6aa898-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.894390] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "c4718797-aa86-4ec0-94d3-6480bd6aa898-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.894538] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "c4718797-aa86-4ec0-94d3-6480bd6aa898-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.896431] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-edafe718-5fac-49ee-a4de-fd758143f31c tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1037.896882] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e856c051-8e88-4b48-82eb-03907bb5b28b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.899314] env[63372]: INFO nova.compute.manager [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Terminating instance [ 1037.902207] env[63372]: DEBUG nova.compute.manager [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1037.902475] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1037.903686] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033bd413-6b0d-4ce9-8270-29fe185faf7d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.933042] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Reconfiguring VM to detach interface {{(pid=63372) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1037.933042] env[63372]: DEBUG oslo_vmware.api [None req-edafe718-5fac-49ee-a4de-fd758143f31c tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1037.933042] env[63372]: value = "task-1024712" [ 1037.933042] env[63372]: _type = "Task" [ 1037.933042] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.933956] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95e6be94-0fba-4c0a-9da7-8bc341e15620 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.953842] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1037.954678] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-265df6d1-8d13-4e4d-93ed-ab218c2db512 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.962382] env[63372]: DEBUG oslo_vmware.api [None req-edafe718-5fac-49ee-a4de-fd758143f31c tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024712, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.964742] env[63372]: DEBUG oslo_vmware.api [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 1037.964742] env[63372]: value = "task-1024713" [ 1037.964742] env[63372]: _type = "Task" [ 1037.964742] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.964992] env[63372]: DEBUG oslo_vmware.api [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 1037.964992] env[63372]: value = "task-1024714" [ 1037.964992] env[63372]: _type = "Task" [ 1037.964992] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.976826] env[63372]: DEBUG oslo_vmware.api [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024713, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.980398] env[63372]: DEBUG oslo_vmware.api [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024714, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.133215] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024711, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.97829} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.133454] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d/83db51f8-2c4b-4ef1-afcb-ecac63bcab9d.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1038.133794] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1038.136439] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-54a0c7c6-4ed1-4ca7-8188-ab7edc95c383 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.146271] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1038.146271] env[63372]: value = "task-1024715" [ 1038.146271] env[63372]: _type = "Task" [ 1038.146271] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.158214] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024715, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.182362] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428b6995-efc1-49ed-bd27-9f6f95190909 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.192219] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714e2218-e6dd-4430-9b18-d989a18d8ebe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.224645] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09bfca50-8ebc-49cc-964e-3f0ac695c79a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.232762] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd7105e-2e72-43ea-b1b7-25be7bf14267 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.247812] env[63372]: DEBUG nova.compute.provider_tree [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.324613] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.324953] env[63372]: DEBUG nova.compute.manager [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Instance network_info: |[{"id": "0fbd29da-197e-452a-9644-ab461d995ec3", "address": "fa:16:3e:c8:e0:b1", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fbd29da-19", "ovs_interfaceid": "0fbd29da-197e-452a-9644-ab461d995ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1038.325406] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:e0:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9f208df-1fb5-4403-9796-7fd19e4bfb85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0fbd29da-197e-452a-9644-ab461d995ec3', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1038.332844] env[63372]: DEBUG oslo.service.loopingcall [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1038.333074] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1038.333309] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cbc05517-fdea-4f07-ab44-58b334666e91 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.353483] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1038.353483] env[63372]: value = "task-1024716" [ 1038.353483] env[63372]: _type = "Task" [ 1038.353483] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.360849] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024716, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.457835] env[63372]: DEBUG oslo_vmware.api [None req-edafe718-5fac-49ee-a4de-fd758143f31c tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024712, 'name': PowerOffVM_Task, 'duration_secs': 0.23794} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.458218] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-edafe718-5fac-49ee-a4de-fd758143f31c tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1038.458337] env[63372]: DEBUG nova.compute.manager [None req-edafe718-5fac-49ee-a4de-fd758143f31c tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.459117] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9fecc2-d4f1-4f4a-bc17-f3483b58bf90 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.478295] env[63372]: DEBUG oslo_vmware.api [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024713, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.481546] env[63372]: DEBUG oslo_vmware.api [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024714, 'name': PowerOffVM_Task, 'duration_secs': 0.205869} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.481796] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1038.482028] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1038.482164] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-788d7ccb-7924-4843-ad76-428933680ec1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.536198] env[63372]: DEBUG nova.compute.manager [req-99a84172-947f-46ee-9536-17bef11f934f req-18ab7026-2958-453a-af2b-8eabc9995581 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Received event network-changed-0fbd29da-197e-452a-9644-ab461d995ec3 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1038.536401] env[63372]: DEBUG nova.compute.manager [req-99a84172-947f-46ee-9536-17bef11f934f req-18ab7026-2958-453a-af2b-8eabc9995581 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Refreshing instance network info cache due to event network-changed-0fbd29da-197e-452a-9644-ab461d995ec3. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1038.536612] env[63372]: DEBUG oslo_concurrency.lockutils [req-99a84172-947f-46ee-9536-17bef11f934f req-18ab7026-2958-453a-af2b-8eabc9995581 service nova] Acquiring lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.536756] env[63372]: DEBUG oslo_concurrency.lockutils [req-99a84172-947f-46ee-9536-17bef11f934f req-18ab7026-2958-453a-af2b-8eabc9995581 service nova] Acquired lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.536916] env[63372]: DEBUG nova.network.neutron [req-99a84172-947f-46ee-9536-17bef11f934f req-18ab7026-2958-453a-af2b-8eabc9995581 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Refreshing network info cache for port 0fbd29da-197e-452a-9644-ab461d995ec3 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1038.650791] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1038.651127] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1038.651335] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleting the datastore file [datastore2] c4718797-aa86-4ec0-94d3-6480bd6aa898 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.651968] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8511330d-1a4b-495c-8d49-21e59fc2598e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.656716] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024715, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093562} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.657288] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1038.658098] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83dbf615-41bb-4312-828a-64accf3452e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.661974] env[63372]: DEBUG oslo_vmware.api [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for the task: (returnval){ [ 1038.661974] env[63372]: value = "task-1024718" [ 1038.661974] env[63372]: _type = "Task" [ 1038.661974] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.682439] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d/83db51f8-2c4b-4ef1-afcb-ecac63bcab9d.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1038.683487] env[63372]: DEBUG nova.network.neutron [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance_info_cache with network_info: [{"id": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "address": "fa:16:3e:28:02:b8", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap128a0686-7c", "ovs_interfaceid": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.685198] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0635ed7-f188-447a-bd1d-e29856e3c090 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.704337] env[63372]: DEBUG oslo_vmware.api [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024718, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.709264] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1038.709264] env[63372]: value = "task-1024719" [ 1038.709264] env[63372]: _type = "Task" [ 1038.709264] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.717379] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024719, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.751457] env[63372]: DEBUG nova.scheduler.client.report [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.864230] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024716, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.978758] env[63372]: DEBUG oslo_vmware.api [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024713, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.979680] env[63372]: DEBUG oslo_concurrency.lockutils [None req-edafe718-5fac-49ee-a4de-fd758143f31c tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c38256cc-215c-4bec-94e5-24a1bbef3e10" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.613s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.171791] env[63372]: DEBUG oslo_vmware.api [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Task: {'id': task-1024718, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.425319} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.174317] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.174512] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1039.174770] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1039.174983] env[63372]: INFO nova.compute.manager [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Took 1.27 seconds to destroy the instance on the hypervisor. [ 1039.175248] env[63372]: DEBUG oslo.service.loopingcall [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.175495] env[63372]: DEBUG nova.compute.manager [-] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1039.175598] env[63372]: DEBUG nova.network.neutron [-] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1039.202771] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.224023] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024719, 'name': ReconfigVM_Task, 'duration_secs': 0.422253} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.224023] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d/83db51f8-2c4b-4ef1-afcb-ecac63bcab9d.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1039.224023] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88d478fa-6883-4081-8a80-8e4323282f93 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.229805] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1039.229805] env[63372]: value = "task-1024720" [ 1039.229805] env[63372]: _type = "Task" [ 1039.229805] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.238814] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024720, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.250667] env[63372]: DEBUG nova.network.neutron [req-99a84172-947f-46ee-9536-17bef11f934f req-18ab7026-2958-453a-af2b-8eabc9995581 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Updated VIF entry in instance network info cache for port 0fbd29da-197e-452a-9644-ab461d995ec3. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1039.251043] env[63372]: DEBUG nova.network.neutron [req-99a84172-947f-46ee-9536-17bef11f934f req-18ab7026-2958-453a-af2b-8eabc9995581 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Updating instance_info_cache with network_info: [{"id": "0fbd29da-197e-452a-9644-ab461d995ec3", "address": "fa:16:3e:c8:e0:b1", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fbd29da-19", "ovs_interfaceid": "0fbd29da-197e-452a-9644-ab461d995ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.256283] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.289s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.283491] env[63372]: INFO nova.network.neutron [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updating port 57943c40-a8c1-4d1a-bb83-d257029ce77f with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1039.365344] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024716, 'name': CreateVM_Task, 'duration_secs': 0.855825} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.365517] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1039.366413] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.366585] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.366928] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1039.367216] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41ccb86b-84c9-4c58-bb21-4ea6056610a3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.373066] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1039.373066] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52e86965-1d3d-f9d8-620a-2b6040b8b3ad" [ 1039.373066] env[63372]: _type = "Task" [ 1039.373066] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.381995] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e86965-1d3d-f9d8-620a-2b6040b8b3ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.477487] env[63372]: DEBUG oslo_vmware.api [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024713, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.706488] env[63372]: DEBUG nova.compute.manager [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63372) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1039.739738] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024720, 'name': Rename_Task, 'duration_secs': 0.168409} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.740221] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1039.740494] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b1c3d52-3399-44bd-ace4-2832960633f4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.746496] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1039.746496] env[63372]: value = "task-1024721" [ 1039.746496] env[63372]: _type = "Task" [ 1039.746496] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.755790] env[63372]: DEBUG oslo_concurrency.lockutils [req-99a84172-947f-46ee-9536-17bef11f934f req-18ab7026-2958-453a-af2b-8eabc9995581 service nova] Releasing lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.756487] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024721, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.883349] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52e86965-1d3d-f9d8-620a-2b6040b8b3ad, 'name': SearchDatastore_Task, 'duration_secs': 0.013789} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.883660] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.883900] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1039.884158] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.884394] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.884617] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1039.884926] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-681172c0-0726-4dcc-b5fc-5013866dbcdf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.896325] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1039.896559] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1039.897402] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41b552c9-f693-4d43-bb45-68f6094699e6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.902732] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1039.902732] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52dc334b-b7dc-4cf5-19e2-347edf3d8b29" [ 1039.902732] env[63372]: _type = "Task" [ 1039.902732] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.910267] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52dc334b-b7dc-4cf5-19e2-347edf3d8b29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.977856] env[63372]: DEBUG oslo_vmware.api [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024713, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.997517] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "c38256cc-215c-4bec-94e5-24a1bbef3e10" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.997799] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c38256cc-215c-4bec-94e5-24a1bbef3e10" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.998191] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "c38256cc-215c-4bec-94e5-24a1bbef3e10-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.998298] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c38256cc-215c-4bec-94e5-24a1bbef3e10-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.998410] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c38256cc-215c-4bec-94e5-24a1bbef3e10-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.000736] env[63372]: INFO nova.compute.manager [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Terminating instance [ 1040.002678] env[63372]: DEBUG nova.compute.manager [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1040.002868] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1040.003714] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56c9a6f-bad9-4a48-8df1-166ee18dab09 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.011082] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1040.011330] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8ce85025-0980-4a3c-b74f-cfea54a3bf2c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.053772] env[63372]: DEBUG nova.network.neutron [-] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.085721] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1040.085774] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1040.085997] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleting the datastore file [datastore2] c38256cc-215c-4bec-94e5-24a1bbef3e10 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1040.086308] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee820640-9001-4717-8781-739d2db953af {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.094026] env[63372]: DEBUG oslo_vmware.api [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1040.094026] env[63372]: value = "task-1024723" [ 1040.094026] env[63372]: _type = "Task" [ 1040.094026] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.101558] env[63372]: DEBUG oslo_vmware.api [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024723, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.256863] env[63372]: DEBUG oslo_vmware.api [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024721, 'name': PowerOnVM_Task, 'duration_secs': 0.46274} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.257148] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1040.257354] env[63372]: INFO nova.compute.manager [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Took 8.72 seconds to spawn the instance on the hypervisor. [ 1040.257533] env[63372]: DEBUG nova.compute.manager [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1040.258272] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cefce57b-9062-4930-946e-1be79186b7cb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.413274] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52dc334b-b7dc-4cf5-19e2-347edf3d8b29, 'name': SearchDatastore_Task, 'duration_secs': 0.018546} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.414115] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b473a15d-962c-4f48-96e7-ed84f5dbc482 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.419056] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1040.419056] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520f44ed-deac-9ded-205d-134fb241ef98" [ 1040.419056] env[63372]: _type = "Task" [ 1040.419056] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.426482] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520f44ed-deac-9ded-205d-134fb241ef98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.478349] env[63372]: DEBUG oslo_vmware.api [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024713, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.556593] env[63372]: INFO nova.compute.manager [-] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Took 1.38 seconds to deallocate network for instance. [ 1040.564688] env[63372]: DEBUG nova.compute.manager [req-cfdde732-842c-40ab-8803-d22c4c4a1125 req-c9103f6b-ae82-44ab-bcb9-33f41e7d6e60 service nova] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Received event network-vif-deleted-fff45461-b9dc-4d60-989d-c7460bc0becb {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.604209] env[63372]: DEBUG oslo_vmware.api [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024723, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.205961} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.604461] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1040.604648] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1040.604827] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1040.604997] env[63372]: INFO nova.compute.manager [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1040.605258] env[63372]: DEBUG oslo.service.loopingcall [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1040.605451] env[63372]: DEBUG nova.compute.manager [-] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1040.605543] env[63372]: DEBUG nova.network.neutron [-] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1040.778664] env[63372]: INFO nova.compute.manager [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Took 14.29 seconds to build instance. [ 1040.817947] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.817947] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.929408] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520f44ed-deac-9ded-205d-134fb241ef98, 'name': SearchDatastore_Task, 'duration_secs': 0.014485} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.929645] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.929905] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] a6bd1c51-e702-48e4-b9ab-8d80db1fec9e/a6bd1c51-e702-48e4-b9ab-8d80db1fec9e.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1040.930182] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c6e955e-1072-43f9-9a2e-674439d40d23 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.936760] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1040.936760] env[63372]: value = "task-1024724" [ 1040.936760] env[63372]: _type = "Task" [ 1040.936760] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.944684] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024724, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.978347] env[63372]: DEBUG oslo_vmware.api [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024713, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.064103] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.281026] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8c837f44-51ee-40c9-8895-aea657fffe57 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.805s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.319856] env[63372]: DEBUG nova.objects.instance [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'migration_context' on Instance uuid b7732621-7e58-40a3-b723-5c66df6f74a1 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.447616] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024724, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458092} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.448009] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] a6bd1c51-e702-48e4-b9ab-8d80db1fec9e/a6bd1c51-e702-48e4-b9ab-8d80db1fec9e.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1041.448266] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1041.448532] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-49133685-1e6c-4126-a147-77cc64e4477e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.454854] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1041.454854] env[63372]: value = "task-1024725" [ 1041.454854] env[63372]: _type = "Task" [ 1041.454854] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.462597] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024725, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.476978] env[63372]: DEBUG oslo_vmware.api [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024713, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.521598] env[63372]: DEBUG nova.network.neutron [-] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.966105] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024725, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067777} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.968817] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1041.970110] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759a0875-1d9d-489f-8ab4-cd924db05de8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.990783] env[63372]: DEBUG oslo_vmware.api [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024713, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.999403] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] a6bd1c51-e702-48e4-b9ab-8d80db1fec9e/a6bd1c51-e702-48e4-b9ab-8d80db1fec9e.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1042.000465] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56816c35-4aea-4cd6-94d5-5e3c542e9613 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.015900] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-253c5917-ca0b-4473-847a-a7b00f7c3f15 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.024790] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0679b12-ac79-4c64-8f93-6c5165a39d1e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.027928] env[63372]: INFO nova.compute.manager [-] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Took 1.42 seconds to deallocate network for instance. [ 1042.028712] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1042.028712] env[63372]: value = "task-1024726" [ 1042.028712] env[63372]: _type = "Task" [ 1042.028712] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.061484] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513eff3f-499b-46a8-9d69-430dd556189c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.067355] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024726, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.074105] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41486cb6-7ca1-46ba-b7cf-5e93380ffe5b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.088202] env[63372]: DEBUG nova.compute.provider_tree [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.259165] env[63372]: DEBUG nova.compute.manager [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Stashing vm_state: active {{(pid=63372) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1042.482340] env[63372]: DEBUG oslo_vmware.api [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024713, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.541180] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024726, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.568800] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.589406] env[63372]: DEBUG nova.compute.manager [req-60c84b3b-2404-4c22-b717-613fc5a50c0c req-328ba5f2-a88f-4107-9c6b-40a80ca6ce15 service nova] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Received event network-vif-deleted-9281c23e-1f88-4862-8f9a-e3e3478592dc {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.592586] env[63372]: DEBUG nova.scheduler.client.report [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1042.781157] env[63372]: DEBUG oslo_concurrency.lockutils [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.983556] env[63372]: DEBUG oslo_vmware.api [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024713, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.039861] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024726, 'name': ReconfigVM_Task, 'duration_secs': 0.747959} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.040348] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Reconfigured VM instance instance-00000068 to attach disk [datastore1] a6bd1c51-e702-48e4-b9ab-8d80db1fec9e/a6bd1c51-e702-48e4-b9ab-8d80db1fec9e.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1043.040999] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-41ab37e7-6a3b-4da7-96aa-349a3824caf5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.047036] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1043.047036] env[63372]: value = "task-1024727" [ 1043.047036] env[63372]: _type = "Task" [ 1043.047036] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.054981] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024727, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.483975] env[63372]: DEBUG oslo_vmware.api [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024713, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.562660] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024727, 'name': Rename_Task, 'duration_secs': 0.172214} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.563030] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1043.563298] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e245bc7-07b9-4c49-a40d-53cac101b04c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.572937] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1043.572937] env[63372]: value = "task-1024728" [ 1043.572937] env[63372]: _type = "Task" [ 1043.572937] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.582123] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024728, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.608340] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.791s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.614312] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.550s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.614628] env[63372]: DEBUG nova.objects.instance [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lazy-loading 'resources' on Instance uuid c4718797-aa86-4ec0-94d3-6480bd6aa898 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.984383] env[63372]: DEBUG oslo_vmware.api [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024713, 'name': ReconfigVM_Task, 'duration_secs': 5.802174} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.984718] env[63372]: DEBUG oslo_concurrency.lockutils [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.984857] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Reconfigured VM to detach interface {{(pid=63372) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1044.084764] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024728, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.326510] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a1a1315-9916-4674-8a24-28bf13af4a93 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.335207] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb06bf94-d487-4a35-b0c4-21080607c529 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.371218] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7024077e-e230-414e-bfec-2d34bb4826db {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.378817] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0dae89-1d71-4771-8278-c23d83e87ab1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.394821] env[63372]: DEBUG nova.compute.provider_tree [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.584828] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024728, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.898099] env[63372]: DEBUG nova.scheduler.client.report [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.084641] env[63372]: DEBUG oslo_vmware.api [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024728, 'name': PowerOnVM_Task, 'duration_secs': 1.340149} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.085116] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1045.085116] env[63372]: INFO nova.compute.manager [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Took 8.86 seconds to spawn the instance on the hypervisor. [ 1045.085261] env[63372]: DEBUG nova.compute.manager [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1045.086032] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653080a7-2436-4be3-b38e-c9ea7b25ed23 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.152663] env[63372]: INFO nova.compute.manager [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Swapping old allocation on dict_keys(['c1b110ca-a185-44c2-ba1c-f05b3b420add']) held by migration 7acff7e3-4df1-4ed7-8106-7ee02c83b666 for instance [ 1045.178450] env[63372]: DEBUG nova.scheduler.client.report [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Overwriting current allocation {'allocations': {'c1b110ca-a185-44c2-ba1c-f05b3b420add': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 148}}, 'project_id': '98581685387a4f1499ae6ed378af982c', 'user_id': '0cc3768e4e8c47a88565a21b7f05ad02', 'consumer_generation': 1} on consumer b7732621-7e58-40a3-b723-5c66df6f74a1 {{(pid=63372) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1045.261630] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.261809] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.262128] env[63372]: DEBUG nova.network.neutron [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1045.389074] env[63372]: DEBUG oslo_concurrency.lockutils [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.389437] env[63372]: DEBUG oslo_concurrency.lockutils [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquired lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.389614] env[63372]: DEBUG nova.network.neutron [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1045.403371] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.789s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.405807] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.837s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.406074] env[63372]: DEBUG nova.objects.instance [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lazy-loading 'resources' on Instance uuid c38256cc-215c-4bec-94e5-24a1bbef3e10 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.421277] env[63372]: INFO nova.scheduler.client.report [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Deleted allocations for instance c4718797-aa86-4ec0-94d3-6480bd6aa898 [ 1045.603633] env[63372]: INFO nova.compute.manager [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Took 15.74 seconds to build instance. [ 1045.829856] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.830122] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.830342] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.830530] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.830702] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.833105] env[63372]: INFO nova.compute.manager [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Terminating instance [ 1045.835033] env[63372]: DEBUG nova.compute.manager [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1045.835238] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1045.836096] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b4848a-7134-41a4-9e6e-35e5b4e287e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.844190] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1045.844456] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29746711-2ba7-4490-a43c-6c5922c4c3df {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.850870] env[63372]: DEBUG oslo_vmware.api [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 1045.850870] env[63372]: value = "task-1024729" [ 1045.850870] env[63372]: _type = "Task" [ 1045.850870] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.861798] env[63372]: DEBUG oslo_vmware.api [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024729, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.929403] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d0f9515-ffcd-45e3-8747-e9c09bdc9863 tempest-ServersTestJSON-580861918 tempest-ServersTestJSON-580861918-project-member] Lock "c4718797-aa86-4ec0-94d3-6480bd6aa898" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.035s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.088495] env[63372]: DEBUG nova.network.neutron [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance_info_cache with network_info: [{"id": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "address": "fa:16:3e:28:02:b8", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap128a0686-7c", "ovs_interfaceid": "128a0686-7ca8-427f-9cf7-c63a2f370e1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.098024] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd1f25d-06f1-47cf-909c-4f847c8934af {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.105403] env[63372]: DEBUG oslo_concurrency.lockutils [None req-852d0943-891e-49d0-a442-f9bf5e50b966 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.254s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.110096] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71cb400c-e8a7-41b1-bb3f-15816f632b66 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.144303] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f63129b-5d46-47cf-8bea-a0cea2f9ec6b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.155667] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce2b8d2-5a9d-4291-ae74-a409da1c1627 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.172427] env[63372]: DEBUG nova.compute.provider_tree [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.230215] env[63372]: INFO nova.network.neutron [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Port 1d17bf13-a03b-4654-bdfe-adc2d47e273a from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1046.230622] env[63372]: DEBUG nova.network.neutron [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Updating instance_info_cache with network_info: [{"id": "301c1926-329c-454a-9755-e33bef4550d8", "address": "fa:16:3e:bc:d9:91", "network": {"id": "e1f7ca1b-541e-468f-a439-a6841fc781ba", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1526482324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46d2cb188a004dc385f2aa15458be929", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap301c1926-32", "ovs_interfaceid": "301c1926-329c-454a-9755-e33bef4550d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.360933] env[63372]: DEBUG oslo_vmware.api [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024729, 'name': PowerOffVM_Task, 'duration_secs': 0.19398} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.361882] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1046.362131] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1046.362382] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce84ba48-b6ae-4866-9c67-d5e620bd18f8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.432308] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1046.432552] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1046.432767] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Deleting the datastore file [datastore2] d7b7ee35-5e83-4c62-bd1e-8ec39951c44c {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1046.433073] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-41688af6-d829-4edb-92f7-58b9baa76801 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.440312] env[63372]: DEBUG oslo_vmware.api [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 1046.440312] env[63372]: value = "task-1024731" [ 1046.440312] env[63372]: _type = "Task" [ 1046.440312] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.448080] env[63372]: DEBUG oslo_vmware.api [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024731, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.578739] env[63372]: DEBUG nova.compute.manager [req-4842a6df-84df-493f-974a-50adca5f9f24 req-dc8fa4fb-fa71-4dba-a02f-a41b8a10e172 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Received event network-changed-0fbd29da-197e-452a-9644-ab461d995ec3 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1046.578739] env[63372]: DEBUG nova.compute.manager [req-4842a6df-84df-493f-974a-50adca5f9f24 req-dc8fa4fb-fa71-4dba-a02f-a41b8a10e172 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Refreshing instance network info cache due to event network-changed-0fbd29da-197e-452a-9644-ab461d995ec3. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1046.578739] env[63372]: DEBUG oslo_concurrency.lockutils [req-4842a6df-84df-493f-974a-50adca5f9f24 req-dc8fa4fb-fa71-4dba-a02f-a41b8a10e172 service nova] Acquiring lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.578739] env[63372]: DEBUG oslo_concurrency.lockutils [req-4842a6df-84df-493f-974a-50adca5f9f24 req-dc8fa4fb-fa71-4dba-a02f-a41b8a10e172 service nova] Acquired lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.578739] env[63372]: DEBUG nova.network.neutron [req-4842a6df-84df-493f-974a-50adca5f9f24 req-dc8fa4fb-fa71-4dba-a02f-a41b8a10e172 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Refreshing network info cache for port 0fbd29da-197e-452a-9644-ab461d995ec3 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1046.591209] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "refresh_cache-b7732621-7e58-40a3-b723-5c66df6f74a1" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.592968] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f63e134-1a07-4ab3-9181-85a8aa38901b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.605098] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934f0f17-7e9d-437c-8b7f-49403f89a6eb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.675702] env[63372]: DEBUG nova.scheduler.client.report [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1046.733524] env[63372]: DEBUG oslo_concurrency.lockutils [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Releasing lock "refresh_cache-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.950383] env[63372]: DEBUG oslo_vmware.api [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024731, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162767} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.950652] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1046.950847] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1046.951042] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1046.951309] env[63372]: INFO nova.compute.manager [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1046.951560] env[63372]: DEBUG oslo.service.loopingcall [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.951763] env[63372]: DEBUG nova.compute.manager [-] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1046.951858] env[63372]: DEBUG nova.network.neutron [-] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1047.181023] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.775s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.184134] env[63372]: DEBUG oslo_concurrency.lockutils [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.402s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.214251] env[63372]: INFO nova.scheduler.client.report [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleted allocations for instance c38256cc-215c-4bec-94e5-24a1bbef3e10 [ 1047.243101] env[63372]: DEBUG oslo_concurrency.lockutils [None req-319f68d3-d46d-41d3-b7f7-6ff3c287adbc tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "interface-d7b7ee35-5e83-4c62-bd1e-8ec39951c44c-1d17bf13-a03b-4654-bdfe-adc2d47e273a" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.880s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.481834] env[63372]: DEBUG nova.network.neutron [req-4842a6df-84df-493f-974a-50adca5f9f24 req-dc8fa4fb-fa71-4dba-a02f-a41b8a10e172 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Updated VIF entry in instance network info cache for port 0fbd29da-197e-452a-9644-ab461d995ec3. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1047.481834] env[63372]: DEBUG nova.network.neutron [req-4842a6df-84df-493f-974a-50adca5f9f24 req-dc8fa4fb-fa71-4dba-a02f-a41b8a10e172 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Updating instance_info_cache with network_info: [{"id": "0fbd29da-197e-452a-9644-ab461d995ec3", "address": "fa:16:3e:c8:e0:b1", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fbd29da-19", "ovs_interfaceid": "0fbd29da-197e-452a-9644-ab461d995ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.691523] env[63372]: INFO nova.compute.claims [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1047.715157] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1047.719152] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59154753-51f1-46b0-84e2-39c64730b160 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.726152] env[63372]: DEBUG oslo_concurrency.lockutils [None req-84dbe752-7842-4969-af1b-d4ef1d0c0649 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c38256cc-215c-4bec-94e5-24a1bbef3e10" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.728s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.729101] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1047.729101] env[63372]: value = "task-1024732" [ 1047.729101] env[63372]: _type = "Task" [ 1047.729101] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.743324] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024732, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.894323] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.894541] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquired lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.894685] env[63372]: DEBUG nova.network.neutron [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1047.984691] env[63372]: DEBUG oslo_concurrency.lockutils [req-4842a6df-84df-493f-974a-50adca5f9f24 req-dc8fa4fb-fa71-4dba-a02f-a41b8a10e172 service nova] Releasing lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.177114] env[63372]: DEBUG nova.network.neutron [-] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.199706] env[63372]: INFO nova.compute.resource_tracker [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating resource usage from migration 0f45e8fc-73e0-4bc6-a3a7-a5ae749ae41d [ 1048.243694] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024732, 'name': PowerOffVM_Task, 'duration_secs': 0.207938} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.247157] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1048.247288] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1048.247512] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1048.247707] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.247915] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1048.248127] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.248418] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1048.248418] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1048.248620] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1048.248922] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1048.248922] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1048.249082] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1048.255180] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d091514e-2708-40f0-93a3-bed72853e99d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.270662] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1048.270662] env[63372]: value = "task-1024733" [ 1048.270662] env[63372]: _type = "Task" [ 1048.270662] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.279305] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024733, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.406052] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3dfe4f5-fdb1-40ce-99fc-c4278634aae7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.418999] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e15df523-c7e5-4dda-aa43-259e6a510a1f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.479409] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de80bf2-7f51-4efe-8071-8f44a9601b41 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.489877] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ef9519-51c6-4725-a056-aacea4d7503c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.501069] env[63372]: DEBUG nova.compute.provider_tree [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.531281] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "1570dee9-33d5-4af9-afe8-37cb4c331113" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.531540] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "1570dee9-33d5-4af9-afe8-37cb4c331113" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.613377] env[63372]: DEBUG nova.compute.manager [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Received event network-vif-deleted-301c1926-329c-454a-9755-e33bef4550d8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1048.613605] env[63372]: DEBUG nova.compute.manager [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Received event network-vif-plugged-57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1048.613792] env[63372]: DEBUG oslo_concurrency.lockutils [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] Acquiring lock "91985614-b959-401e-bb06-d67b230ee026-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.614018] env[63372]: DEBUG oslo_concurrency.lockutils [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] Lock "91985614-b959-401e-bb06-d67b230ee026-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.614270] env[63372]: DEBUG oslo_concurrency.lockutils [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] Lock "91985614-b959-401e-bb06-d67b230ee026-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.614445] env[63372]: DEBUG nova.compute.manager [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] No waiting events found dispatching network-vif-plugged-57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1048.614702] env[63372]: WARNING nova.compute.manager [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Received unexpected event network-vif-plugged-57943c40-a8c1-4d1a-bb83-d257029ce77f for instance with vm_state shelved_offloaded and task_state spawning. [ 1048.614934] env[63372]: DEBUG nova.compute.manager [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Received event network-changed-57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1048.615042] env[63372]: DEBUG nova.compute.manager [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Refreshing instance network info cache due to event network-changed-57943c40-a8c1-4d1a-bb83-d257029ce77f. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1048.615216] env[63372]: DEBUG oslo_concurrency.lockutils [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] Acquiring lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.681142] env[63372]: INFO nova.compute.manager [-] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Took 1.73 seconds to deallocate network for instance. [ 1048.754276] env[63372]: DEBUG nova.network.neutron [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updating instance_info_cache with network_info: [{"id": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "address": "fa:16:3e:1f:74:b1", "network": {"id": "cbdcb945-f0a5-49b1-9b18-fcd8e4ca9728", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1068757189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e517915d5e7a4e3e924cc95c55170a66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57943c40-a8", "ovs_interfaceid": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.783090] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024733, 'name': ReconfigVM_Task, 'duration_secs': 0.153361} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.784084] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d67b13-ea44-4276-8ea3-44453f62edab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.805369] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1048.805696] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1048.805793] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.805975] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1048.806137] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.806292] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1048.806494] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1048.806657] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1048.806825] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1048.806990] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1048.807194] env[63372]: DEBUG nova.virt.hardware [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1048.808045] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d2b475a-5bf2-40e2-b144-ef789c0d373e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.815037] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1048.815037] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]522471b2-fb32-cb9b-ed87-d013e0634b85" [ 1048.815037] env[63372]: _type = "Task" [ 1048.815037] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.823863] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522471b2-fb32-cb9b-ed87-d013e0634b85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.005026] env[63372]: DEBUG nova.scheduler.client.report [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.034412] env[63372]: DEBUG nova.compute.manager [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1049.188130] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.257615] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Releasing lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.260089] env[63372]: DEBUG oslo_concurrency.lockutils [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] Acquired lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.262429] env[63372]: DEBUG nova.network.neutron [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Refreshing network info cache for port 57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1049.284654] env[63372]: DEBUG nova.virt.hardware [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='25fd2967e1a76d2b63582696b82ac74e',container_format='bare',created_at=2024-09-30T11:42:38Z,direct_url=,disk_format='vmdk',id=003d2e97-266b-4459-abd7-6ca82ed4825e,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-398773629-shelved',owner='e517915d5e7a4e3e924cc95c55170a66',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-09-30T11:42:58Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1049.284654] env[63372]: DEBUG nova.virt.hardware [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1049.284886] env[63372]: DEBUG nova.virt.hardware [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1049.284886] env[63372]: DEBUG nova.virt.hardware [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1049.285016] env[63372]: DEBUG nova.virt.hardware [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1049.285189] env[63372]: DEBUG nova.virt.hardware [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1049.285406] env[63372]: DEBUG nova.virt.hardware [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1049.285570] env[63372]: DEBUG nova.virt.hardware [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1049.286082] env[63372]: DEBUG nova.virt.hardware [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1049.286082] env[63372]: DEBUG nova.virt.hardware [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1049.286082] env[63372]: DEBUG nova.virt.hardware [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1049.286954] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31796758-7c2b-4fe3-ade2-9f7907536b17 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.295498] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7095089b-31e5-4399-9f5b-39aad98a059a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.308888] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:74:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd3c6b64-aba2-4bdc-a693-3b4dff3ed861', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '57943c40-a8c1-4d1a-bb83-d257029ce77f', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1049.316365] env[63372]: DEBUG oslo.service.loopingcall [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1049.316711] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91985614-b959-401e-bb06-d67b230ee026] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1049.319488] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-65903530-f702-45f7-b075-8270e9d55303 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.339036] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]522471b2-fb32-cb9b-ed87-d013e0634b85, 'name': SearchDatastore_Task, 'duration_secs': 0.011796} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.345518] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Reconfiguring VM instance instance-0000005a to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1049.345839] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1049.345839] env[63372]: value = "task-1024734" [ 1049.345839] env[63372]: _type = "Task" [ 1049.345839] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.346374] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c186f903-58e2-40d0-9337-ca77af0315ec {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.367090] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024734, 'name': CreateVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.368368] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1049.368368] env[63372]: value = "task-1024735" [ 1049.368368] env[63372]: _type = "Task" [ 1049.368368] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.376271] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024735, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.510595] env[63372]: DEBUG oslo_concurrency.lockutils [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.327s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.511021] env[63372]: INFO nova.compute.manager [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Migrating [ 1049.519844] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.332s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.520088] env[63372]: DEBUG nova.objects.instance [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'resources' on Instance uuid d7b7ee35-5e83-4c62-bd1e-8ec39951c44c {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.559268] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.872806] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024734, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.881415] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024735, 'name': ReconfigVM_Task, 'duration_secs': 0.208787} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.881658] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Reconfigured VM instance instance-0000005a to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1049.882335] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe18b49-819b-4b5c-a1de-52bd1c039f3f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.908824] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] b7732621-7e58-40a3-b723-5c66df6f74a1/b7732621-7e58-40a3-b723-5c66df6f74a1.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.911442] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1733bad8-616a-478e-807c-deb10105d453 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.933954] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1049.933954] env[63372]: value = "task-1024736" [ 1049.933954] env[63372]: _type = "Task" [ 1049.933954] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.943963] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024736, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.031148] env[63372]: DEBUG oslo_concurrency.lockutils [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.031378] env[63372]: DEBUG oslo_concurrency.lockutils [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.031577] env[63372]: DEBUG nova.network.neutron [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1050.087162] env[63372]: DEBUG nova.network.neutron [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updated VIF entry in instance network info cache for port 57943c40-a8c1-4d1a-bb83-d257029ce77f. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1050.087514] env[63372]: DEBUG nova.network.neutron [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updating instance_info_cache with network_info: [{"id": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "address": "fa:16:3e:1f:74:b1", "network": {"id": "cbdcb945-f0a5-49b1-9b18-fcd8e4ca9728", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1068757189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e517915d5e7a4e3e924cc95c55170a66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57943c40-a8", "ovs_interfaceid": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.201428] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c387f491-5b11-46c6-8a8b-ff111c5757bc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.209314] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8ebc44-2da8-4550-9639-1573a81f0eaa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.239612] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1824f0-05b2-41e9-bd8c-9397929270b7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.247150] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e86af2b-0d16-41c3-aa19-4386262789a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.260525] env[63372]: DEBUG nova.compute.provider_tree [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1050.369615] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024734, 'name': CreateVM_Task, 'duration_secs': 0.541132} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.369804] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91985614-b959-401e-bb06-d67b230ee026] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1050.370680] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/003d2e97-266b-4459-abd7-6ca82ed4825e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.370803] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/003d2e97-266b-4459-abd7-6ca82ed4825e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.371245] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/003d2e97-266b-4459-abd7-6ca82ed4825e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1050.371499] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-580be033-1f84-4f13-ac90-1f59a99c646a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.376387] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1050.376387] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5207b832-b164-aa7a-df94-7bf48199fb03" [ 1050.376387] env[63372]: _type = "Task" [ 1050.376387] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.385130] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5207b832-b164-aa7a-df94-7bf48199fb03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.444155] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024736, 'name': ReconfigVM_Task, 'duration_secs': 0.3007} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.444155] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Reconfigured VM instance instance-0000005a to attach disk [datastore2] b7732621-7e58-40a3-b723-5c66df6f74a1/b7732621-7e58-40a3-b723-5c66df6f74a1.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.444729] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae70b50-1cdf-41d8-ae13-591f05f2c1f6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.465122] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1128eeb-1471-4811-a11a-6a1774ac0508 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.485978] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624aa2ce-bcdb-4bb8-be9d-3f5c9dcb3331 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.506969] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5002429a-fa03-4170-bdc0-e12647df4776 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.513908] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1050.514166] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10853ab9-be93-4ac5-9f2f-4b6f42a2737c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.521708] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1050.521708] env[63372]: value = "task-1024737" [ 1050.521708] env[63372]: _type = "Task" [ 1050.521708] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.534424] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024737, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.593751] env[63372]: DEBUG oslo_concurrency.lockutils [req-5443cdeb-82c6-4d64-b1ee-a3ad701d0495 req-c0e95788-041a-414b-bfae-5e645e6a20cf service nova] Releasing lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.734161] env[63372]: DEBUG nova.network.neutron [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating instance_info_cache with network_info: [{"id": "ebc3af7a-29a8-44e6-85e7-4ae1bafa1844", "address": "fa:16:3e:05:a3:85", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebc3af7a-29", "ovs_interfaceid": "ebc3af7a-29a8-44e6-85e7-4ae1bafa1844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.763522] env[63372]: DEBUG nova.scheduler.client.report [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1050.887260] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/003d2e97-266b-4459-abd7-6ca82ed4825e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.887684] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Processing image 003d2e97-266b-4459-abd7-6ca82ed4825e {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1050.888070] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/003d2e97-266b-4459-abd7-6ca82ed4825e/003d2e97-266b-4459-abd7-6ca82ed4825e.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.888324] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/003d2e97-266b-4459-abd7-6ca82ed4825e/003d2e97-266b-4459-abd7-6ca82ed4825e.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.888602] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1050.888936] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1441cb7-7975-4b9e-bf9e-6c722e370bb7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.913035] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1050.913330] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1050.914442] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b5b8b00-55eb-460d-ac41-63f404007852 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.920856] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1050.920856] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5297164e-dc93-adef-7afa-75355352ee88" [ 1050.920856] env[63372]: _type = "Task" [ 1050.920856] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.931607] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5297164e-dc93-adef-7afa-75355352ee88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.033521] env[63372]: DEBUG oslo_vmware.api [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024737, 'name': PowerOnVM_Task, 'duration_secs': 0.378545} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.033980] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1051.237286] env[63372]: DEBUG oslo_concurrency.lockutils [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.268556] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.749s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.271353] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.712s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.272994] env[63372]: INFO nova.compute.claims [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1051.295564] env[63372]: INFO nova.scheduler.client.report [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Deleted allocations for instance d7b7ee35-5e83-4c62-bd1e-8ec39951c44c [ 1051.430773] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Preparing fetch location {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1051.431067] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Fetch image to [datastore2] OSTACK_IMG_5d8bde80-5520-473a-b3b0-c87445e752ad/OSTACK_IMG_5d8bde80-5520-473a-b3b0-c87445e752ad.vmdk {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1051.431293] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Downloading stream optimized image 003d2e97-266b-4459-abd7-6ca82ed4825e to [datastore2] OSTACK_IMG_5d8bde80-5520-473a-b3b0-c87445e752ad/OSTACK_IMG_5d8bde80-5520-473a-b3b0-c87445e752ad.vmdk on the data store datastore2 as vApp {{(pid=63372) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1051.431478] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Downloading image file data 003d2e97-266b-4459-abd7-6ca82ed4825e to the ESX as VM named 'OSTACK_IMG_5d8bde80-5520-473a-b3b0-c87445e752ad' {{(pid=63372) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1051.506692] env[63372]: DEBUG oslo_vmware.rw_handles [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1051.506692] env[63372]: value = "resgroup-9" [ 1051.506692] env[63372]: _type = "ResourcePool" [ 1051.506692] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1051.507231] env[63372]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-035b367c-29d6-42a5-91b6-59531898f1d0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.530106] env[63372]: DEBUG oslo_vmware.rw_handles [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lease: (returnval){ [ 1051.530106] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529adafe-7930-adb8-5a40-5330c53608a5" [ 1051.530106] env[63372]: _type = "HttpNfcLease" [ 1051.530106] env[63372]: } obtained for vApp import into resource pool (val){ [ 1051.530106] env[63372]: value = "resgroup-9" [ 1051.530106] env[63372]: _type = "ResourcePool" [ 1051.530106] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1051.530358] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the lease: (returnval){ [ 1051.530358] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529adafe-7930-adb8-5a40-5330c53608a5" [ 1051.530358] env[63372]: _type = "HttpNfcLease" [ 1051.530358] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1051.536484] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1051.536484] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529adafe-7930-adb8-5a40-5330c53608a5" [ 1051.536484] env[63372]: _type = "HttpNfcLease" [ 1051.536484] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1051.803775] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ce65c1fe-0b5d-4903-b341-d056a8054b3e tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "d7b7ee35-5e83-4c62-bd1e-8ec39951c44c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.974s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.038763] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1052.038763] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529adafe-7930-adb8-5a40-5330c53608a5" [ 1052.038763] env[63372]: _type = "HttpNfcLease" [ 1052.038763] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1052.084804] env[63372]: DEBUG oslo_concurrency.lockutils [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "28eb1675-b1b8-46a6-873d-5c858b716575" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.085116] env[63372]: DEBUG oslo_concurrency.lockutils [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "28eb1675-b1b8-46a6-873d-5c858b716575" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.085258] env[63372]: INFO nova.compute.manager [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Shelving [ 1052.089871] env[63372]: INFO nova.compute.manager [None req-84971bf3-0fb0-4148-893d-88dba42f0d67 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance to original state: 'active' [ 1052.424700] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9462528-97cb-42ec-8482-e914b0d95705 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.431903] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efad4339-0008-4f9b-8f4a-86c8fa101da0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.461761] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-091faca0-146f-4d2c-8379-ceea690a142f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.468507] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d841091d-bfb2-40fd-b3f5-4a818c3ee3b0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.481294] env[63372]: DEBUG nova.compute.provider_tree [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1052.539059] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1052.539059] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529adafe-7930-adb8-5a40-5330c53608a5" [ 1052.539059] env[63372]: _type = "HttpNfcLease" [ 1052.539059] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1052.539324] env[63372]: DEBUG oslo_vmware.rw_handles [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1052.539324] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529adafe-7930-adb8-5a40-5330c53608a5" [ 1052.539324] env[63372]: _type = "HttpNfcLease" [ 1052.539324] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1052.540026] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50578262-a87a-4fee-9216-dc923a294095 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.546966] env[63372]: DEBUG oslo_vmware.rw_handles [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52217b86-b069-01a7-80e1-9f968fb6efbb/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1052.547164] env[63372]: DEBUG oslo_vmware.rw_handles [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52217b86-b069-01a7-80e1-9f968fb6efbb/disk-0.vmdk. {{(pid=63372) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1052.603276] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "b455b16b-3332-43bf-a91e-a350287ba5f3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.603549] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "b455b16b-3332-43bf-a91e-a350287ba5f3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.603757] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "b455b16b-3332-43bf-a91e-a350287ba5f3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.603944] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "b455b16b-3332-43bf-a91e-a350287ba5f3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.604164] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "b455b16b-3332-43bf-a91e-a350287ba5f3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.610893] env[63372]: INFO nova.compute.manager [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Terminating instance [ 1052.615311] env[63372]: DEBUG nova.compute.manager [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1052.615511] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1052.621056] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cbca450-4fd9-480c-82dd-66d09514dcb4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.624358] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1052.625603] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1c208d49-1ad0-4e8e-ab88-f67f3fcb8de5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.627466] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00e9d3f1-a595-4a55-b4ae-a116b612560d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.633968] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1052.636091] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-371fdaf5-d46a-4a68-95e1-44df976b7f8d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.637608] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1052.637608] env[63372]: value = "task-1024739" [ 1052.637608] env[63372]: _type = "Task" [ 1052.637608] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.644112] env[63372]: DEBUG oslo_vmware.api [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 1052.644112] env[63372]: value = "task-1024740" [ 1052.644112] env[63372]: _type = "Task" [ 1052.644112] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.654431] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024739, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.659646] env[63372]: DEBUG oslo_vmware.api [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024740, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.755022] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22267d6-f2d2-439d-9de4-7790c52eb15d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.776035] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating instance '83db51f8-2c4b-4ef1-afcb-ecac63bcab9d' progress to 0 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1052.984981] env[63372]: DEBUG nova.scheduler.client.report [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1053.147514] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024739, 'name': PowerOffVM_Task, 'duration_secs': 0.169493} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.150714] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1053.151571] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4efe3d-fbed-4c31-a95b-c8ab9fc87342 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.174031] env[63372]: DEBUG oslo_vmware.api [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024740, 'name': PowerOffVM_Task, 'duration_secs': 0.187036} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.175392] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1053.175654] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1053.176434] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407c8dea-f170-48f5-9fa7-eae0256dc05f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.179275] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-30d810cc-e8d1-4fd2-947c-d2e65981f1fc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.256466] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1053.256723] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1053.256907] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Deleting the datastore file [datastore2] b455b16b-3332-43bf-a91e-a350287ba5f3 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1053.257195] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-34828e42-e19c-4dfe-a3fc-b8ba90b89f75 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.267074] env[63372]: DEBUG oslo_vmware.api [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for the task: (returnval){ [ 1053.267074] env[63372]: value = "task-1024742" [ 1053.267074] env[63372]: _type = "Task" [ 1053.267074] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.276395] env[63372]: DEBUG oslo_vmware.api [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024742, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.285470] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1053.285906] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "b7732621-7e58-40a3-b723-5c66df6f74a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.286318] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "b7732621-7e58-40a3-b723-5c66df6f74a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.286638] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "b7732621-7e58-40a3-b723-5c66df6f74a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.286888] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "b7732621-7e58-40a3-b723-5c66df6f74a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.287119] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "b7732621-7e58-40a3-b723-5c66df6f74a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.290718] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fe79420-50b5-4f91-adf3-f2f76315a96d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.293147] env[63372]: INFO nova.compute.manager [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Terminating instance [ 1053.296156] env[63372]: DEBUG nova.compute.manager [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1053.296465] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1053.296769] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a90bc6d-5c47-49ce-9286-d7680bc2ba02 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.300379] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1053.300379] env[63372]: value = "task-1024743" [ 1053.300379] env[63372]: _type = "Task" [ 1053.300379] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.307649] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1053.307649] env[63372]: value = "task-1024744" [ 1053.307649] env[63372]: _type = "Task" [ 1053.307649] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.317121] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024743, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.324093] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024744, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.495217] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.224s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.495927] env[63372]: DEBUG nova.compute.manager [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1053.690069] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1053.690449] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4b1d5b52-f681-46cd-8ffd-8b6983ade81f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.699356] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1053.699356] env[63372]: value = "task-1024745" [ 1053.699356] env[63372]: _type = "Task" [ 1053.699356] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.711228] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024745, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.777699] env[63372]: DEBUG oslo_vmware.api [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Task: {'id': task-1024742, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187697} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.779799] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1053.780025] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1053.780214] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1053.780399] env[63372]: INFO nova.compute.manager [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1053.780655] env[63372]: DEBUG oslo.service.loopingcall [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.780859] env[63372]: DEBUG nova.compute.manager [-] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1053.780953] env[63372]: DEBUG nova.network.neutron [-] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1053.815521] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024743, 'name': PowerOffVM_Task, 'duration_secs': 0.20854} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.816256] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1053.816487] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating instance '83db51f8-2c4b-4ef1-afcb-ecac63bcab9d' progress to 17 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1053.826613] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024744, 'name': PowerOffVM_Task, 'duration_secs': 0.217874} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.826894] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1053.827171] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Volume detach. Driver type: vmdk {{(pid=63372) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1053.827402] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227445', 'volume_id': '2d01e3c9-8f3a-4399-a8a8-213b5e212d9a', 'name': 'volume-2d01e3c9-8f3a-4399-a8a8-213b5e212d9a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'b7732621-7e58-40a3-b723-5c66df6f74a1', 'attached_at': '2024-09-30T11:43:23.000000', 'detached_at': '', 'volume_id': '2d01e3c9-8f3a-4399-a8a8-213b5e212d9a', 'serial': '2d01e3c9-8f3a-4399-a8a8-213b5e212d9a'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1053.828290] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8fee853-b3d2-46c8-9f49-6112dc90a49c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.852725] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33159666-35e6-4b13-9bc7-6a9b3c1a4a2c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.860431] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05adc937-fb14-47ea-8713-1ec60231a50e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.883165] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a85a6ae-c2ad-4a76-a745-adfcb7eb8c0d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.899552] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] The volume has not been displaced from its original location: [datastore1] volume-2d01e3c9-8f3a-4399-a8a8-213b5e212d9a/volume-2d01e3c9-8f3a-4399-a8a8-213b5e212d9a.vmdk. No consolidation needed. {{(pid=63372) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1053.908149] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Reconfiguring VM instance instance-0000005a to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1053.908149] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a5de951-df00-4e03-baca-eb823c728e53 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.922401] env[63372]: DEBUG oslo_vmware.rw_handles [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Completed reading data from the image iterator. {{(pid=63372) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1053.922619] env[63372]: DEBUG oslo_vmware.rw_handles [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52217b86-b069-01a7-80e1-9f968fb6efbb/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1053.923847] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3fde125-cd8c-44f6-b7c9-b61b7fdcd4ec {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.930461] env[63372]: DEBUG oslo_vmware.rw_handles [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52217b86-b069-01a7-80e1-9f968fb6efbb/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1053.930675] env[63372]: DEBUG oslo_vmware.rw_handles [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52217b86-b069-01a7-80e1-9f968fb6efbb/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1053.932884] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-142eac96-caf6-4fe1-b702-5a0e09255997 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.937023] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1053.937023] env[63372]: value = "task-1024746" [ 1053.937023] env[63372]: _type = "Task" [ 1053.937023] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.942841] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024746, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.003815] env[63372]: DEBUG nova.compute.utils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1054.005350] env[63372]: DEBUG nova.compute.manager [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1054.005601] env[63372]: DEBUG nova.network.neutron [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1054.089205] env[63372]: DEBUG nova.policy [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '118d2f1de2be45cc8bb48bb75525d37e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'afe39cfedf214d50be775ef736f94da9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1054.212791] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024745, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.247416] env[63372]: DEBUG nova.compute.manager [req-5a3dd1a6-278c-43bd-ad97-7e0c0531fa0b req-694e2c51-0efb-473b-a18e-a36244de6ca9 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Received event network-vif-deleted-18c7850a-d97e-4564-92e5-38d1204369e4 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1054.247617] env[63372]: INFO nova.compute.manager [req-5a3dd1a6-278c-43bd-ad97-7e0c0531fa0b req-694e2c51-0efb-473b-a18e-a36244de6ca9 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Neutron deleted interface 18c7850a-d97e-4564-92e5-38d1204369e4; detaching it from the instance and deleting it from the info cache [ 1054.247838] env[63372]: DEBUG nova.network.neutron [req-5a3dd1a6-278c-43bd-ad97-7e0c0531fa0b req-694e2c51-0efb-473b-a18e-a36244de6ca9 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.323662] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1054.323926] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1054.324105] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1054.324304] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1054.324456] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.324917] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1054.324917] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1054.325035] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1054.325422] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1054.325422] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1054.325568] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1054.331065] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23fc1303-6100-4e8f-92b0-0f572c0b7456 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.347010] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1054.347010] env[63372]: value = "task-1024747" [ 1054.347010] env[63372]: _type = "Task" [ 1054.347010] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.355552] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024747, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.390455] env[63372]: DEBUG nova.network.neutron [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Successfully created port: 6cfbc521-36cc-4af3-b381-f33b3af41bfd {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1054.426060] env[63372]: DEBUG oslo_vmware.rw_handles [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52217b86-b069-01a7-80e1-9f968fb6efbb/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1054.426324] env[63372]: INFO nova.virt.vmwareapi.images [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Downloaded image file data 003d2e97-266b-4459-abd7-6ca82ed4825e [ 1054.427160] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5124bf-4e36-472c-b887-a082fb442927 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.446656] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10a77459-6c29-45b4-bc98-469ab1153d3e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.454787] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024746, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.468939] env[63372]: INFO nova.virt.vmwareapi.images [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] The imported VM was unregistered [ 1054.471288] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Caching image {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1054.471545] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Creating directory with path [datastore2] devstack-image-cache_base/003d2e97-266b-4459-abd7-6ca82ed4825e {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1054.471822] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-72ff276f-cbac-424b-90f0-84d1ce93bd31 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.482476] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Created directory with path [datastore2] devstack-image-cache_base/003d2e97-266b-4459-abd7-6ca82ed4825e {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1054.482677] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_5d8bde80-5520-473a-b3b0-c87445e752ad/OSTACK_IMG_5d8bde80-5520-473a-b3b0-c87445e752ad.vmdk to [datastore2] devstack-image-cache_base/003d2e97-266b-4459-abd7-6ca82ed4825e/003d2e97-266b-4459-abd7-6ca82ed4825e.vmdk. {{(pid=63372) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1054.483119] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-0ad3d352-226b-452d-890f-72189afa7c88 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.489294] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1054.489294] env[63372]: value = "task-1024749" [ 1054.489294] env[63372]: _type = "Task" [ 1054.489294] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.498141] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024749, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.508476] env[63372]: DEBUG nova.compute.manager [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1054.712993] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024745, 'name': CreateSnapshot_Task, 'duration_secs': 0.981299} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.713390] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1054.714191] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1131ddf-f850-41a6-9244-a6456bc76474 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.726128] env[63372]: DEBUG nova.network.neutron [-] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.750326] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-60f123fb-3a4d-4952-adbe-c40cb8a677bb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.760100] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64b5346a-6b59-4ed9-b7cc-7fbd0dc089f1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.788794] env[63372]: DEBUG nova.compute.manager [req-5a3dd1a6-278c-43bd-ad97-7e0c0531fa0b req-694e2c51-0efb-473b-a18e-a36244de6ca9 service nova] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Detach interface failed, port_id=18c7850a-d97e-4564-92e5-38d1204369e4, reason: Instance b455b16b-3332-43bf-a91e-a350287ba5f3 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1054.856922] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024747, 'name': ReconfigVM_Task, 'duration_secs': 0.26706} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.857305] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating instance '83db51f8-2c4b-4ef1-afcb-ecac63bcab9d' progress to 33 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1054.953583] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024746, 'name': ReconfigVM_Task, 'duration_secs': 0.528894} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.953878] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Reconfigured VM instance instance-0000005a to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1054.958637] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb1f4bd7-d2b1-43db-8565-4b057c544494 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.974135] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1054.974135] env[63372]: value = "task-1024750" [ 1054.974135] env[63372]: _type = "Task" [ 1054.974135] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.988358] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024750, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.996290] env[63372]: DEBUG nova.compute.manager [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Stashing vm_state: active {{(pid=63372) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1055.004799] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024749, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.233792] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1055.234326] env[63372]: INFO nova.compute.manager [-] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Took 1.45 seconds to deallocate network for instance. [ 1055.234648] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-833dbf92-1dd3-4522-b0d3-16cf26176490 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.244778] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1055.244778] env[63372]: value = "task-1024751" [ 1055.244778] env[63372]: _type = "Task" [ 1055.244778] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.255182] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024751, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.364578] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1055.364821] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1055.364887] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1055.365084] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1055.365263] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1055.365482] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1055.365699] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1055.365864] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1055.366069] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1055.366251] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1055.366433] env[63372]: DEBUG nova.virt.hardware [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1055.373844] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1055.374258] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bdf7cec8-8c17-4cb6-82ad-52c56f0b7c14 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.396578] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1055.396578] env[63372]: value = "task-1024752" [ 1055.396578] env[63372]: _type = "Task" [ 1055.396578] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.409384] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024752, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.484329] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.506017] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024749, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.518199] env[63372]: DEBUG nova.compute.manager [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1055.522360] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.522591] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.551480] env[63372]: DEBUG nova.virt.hardware [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1055.551751] env[63372]: DEBUG nova.virt.hardware [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1055.552029] env[63372]: DEBUG nova.virt.hardware [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1055.552136] env[63372]: DEBUG nova.virt.hardware [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1055.552296] env[63372]: DEBUG nova.virt.hardware [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1055.552461] env[63372]: DEBUG nova.virt.hardware [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1055.552675] env[63372]: DEBUG nova.virt.hardware [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1055.552839] env[63372]: DEBUG nova.virt.hardware [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1055.553016] env[63372]: DEBUG nova.virt.hardware [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1055.553192] env[63372]: DEBUG nova.virt.hardware [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1055.553369] env[63372]: DEBUG nova.virt.hardware [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1055.554607] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053882b3-3c1b-4305-a786-d165f760f872 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.563084] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77b62771-479a-4822-a788-36c1821d5c05 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.745885] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.757885] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024751, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.907603] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024752, 'name': ReconfigVM_Task, 'duration_secs': 0.245634} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.907920] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1055.908841] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cdb97f4-2f98-43c1-bf64-cb6ec53cecc9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.935638] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d/83db51f8-2c4b-4ef1-afcb-ecac63bcab9d.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1055.937966] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e66f6de7-0519-4946-b0f4-cb1e49200e5e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.956800] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1055.956800] env[63372]: value = "task-1024753" [ 1055.956800] env[63372]: _type = "Task" [ 1055.956800] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.965260] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024753, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.984711] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024750, 'name': ReconfigVM_Task, 'duration_secs': 0.966665} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.984711] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227445', 'volume_id': '2d01e3c9-8f3a-4399-a8a8-213b5e212d9a', 'name': 'volume-2d01e3c9-8f3a-4399-a8a8-213b5e212d9a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'b7732621-7e58-40a3-b723-5c66df6f74a1', 'attached_at': '2024-09-30T11:43:23.000000', 'detached_at': '', 'volume_id': '2d01e3c9-8f3a-4399-a8a8-213b5e212d9a', 'serial': '2d01e3c9-8f3a-4399-a8a8-213b5e212d9a'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1055.985021] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1055.985874] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64270325-f4b4-44a9-b33a-e37af2b4f0c6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.991526] env[63372]: DEBUG nova.compute.manager [req-dd23df82-b514-4809-bb92-bd2b049e527f req-5d3496f0-db97-489a-ae1b-0eed8c72b462 service nova] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Received event network-vif-plugged-6cfbc521-36cc-4af3-b381-f33b3af41bfd {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.991819] env[63372]: DEBUG oslo_concurrency.lockutils [req-dd23df82-b514-4809-bb92-bd2b049e527f req-5d3496f0-db97-489a-ae1b-0eed8c72b462 service nova] Acquiring lock "1570dee9-33d5-4af9-afe8-37cb4c331113-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.992010] env[63372]: DEBUG oslo_concurrency.lockutils [req-dd23df82-b514-4809-bb92-bd2b049e527f req-5d3496f0-db97-489a-ae1b-0eed8c72b462 service nova] Lock "1570dee9-33d5-4af9-afe8-37cb4c331113-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.992290] env[63372]: DEBUG oslo_concurrency.lockutils [req-dd23df82-b514-4809-bb92-bd2b049e527f req-5d3496f0-db97-489a-ae1b-0eed8c72b462 service nova] Lock "1570dee9-33d5-4af9-afe8-37cb4c331113-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.992526] env[63372]: DEBUG nova.compute.manager [req-dd23df82-b514-4809-bb92-bd2b049e527f req-5d3496f0-db97-489a-ae1b-0eed8c72b462 service nova] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] No waiting events found dispatching network-vif-plugged-6cfbc521-36cc-4af3-b381-f33b3af41bfd {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1055.992597] env[63372]: WARNING nova.compute.manager [req-dd23df82-b514-4809-bb92-bd2b049e527f req-5d3496f0-db97-489a-ae1b-0eed8c72b462 service nova] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Received unexpected event network-vif-plugged-6cfbc521-36cc-4af3-b381-f33b3af41bfd for instance with vm_state building and task_state spawning. [ 1055.995703] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1055.999511] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1c726970-8337-42f3-8dad-1a57823276cd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.006557] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024749, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.028238] env[63372]: INFO nova.compute.claims [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1056.138168] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1056.138533] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1056.138899] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleting the datastore file [datastore2] b7732621-7e58-40a3-b723-5c66df6f74a1 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1056.139290] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0de4af3e-b284-414d-a133-a1189fe8a1ba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.149624] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1056.149624] env[63372]: value = "task-1024755" [ 1056.149624] env[63372]: _type = "Task" [ 1056.149624] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.163013] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024755, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.260303] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024751, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.468368] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024753, 'name': ReconfigVM_Task, 'duration_secs': 0.389284} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.468702] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d/83db51f8-2c4b-4ef1-afcb-ecac63bcab9d.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1056.468997] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating instance '83db51f8-2c4b-4ef1-afcb-ecac63bcab9d' progress to 50 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1056.509454] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024749, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.535583] env[63372]: INFO nova.compute.resource_tracker [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating resource usage from migration efaa8b7b-e3cc-47f7-9e4c-0c13bcff933d [ 1056.556342] env[63372]: DEBUG nova.network.neutron [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Successfully updated port: 6cfbc521-36cc-4af3-b381-f33b3af41bfd {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1056.580403] env[63372]: DEBUG nova.compute.manager [req-3b37b469-fb81-44ae-bc09-36672b535815 req-4ab7b04d-138b-4214-8f43-d7b92ad80e6f service nova] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Received event network-changed-6cfbc521-36cc-4af3-b381-f33b3af41bfd {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1056.580693] env[63372]: DEBUG nova.compute.manager [req-3b37b469-fb81-44ae-bc09-36672b535815 req-4ab7b04d-138b-4214-8f43-d7b92ad80e6f service nova] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Refreshing instance network info cache due to event network-changed-6cfbc521-36cc-4af3-b381-f33b3af41bfd. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1056.580841] env[63372]: DEBUG oslo_concurrency.lockutils [req-3b37b469-fb81-44ae-bc09-36672b535815 req-4ab7b04d-138b-4214-8f43-d7b92ad80e6f service nova] Acquiring lock "refresh_cache-1570dee9-33d5-4af9-afe8-37cb4c331113" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.580987] env[63372]: DEBUG oslo_concurrency.lockutils [req-3b37b469-fb81-44ae-bc09-36672b535815 req-4ab7b04d-138b-4214-8f43-d7b92ad80e6f service nova] Acquired lock "refresh_cache-1570dee9-33d5-4af9-afe8-37cb4c331113" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.581221] env[63372]: DEBUG nova.network.neutron [req-3b37b469-fb81-44ae-bc09-36672b535815 req-4ab7b04d-138b-4214-8f43-d7b92ad80e6f service nova] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Refreshing network info cache for port 6cfbc521-36cc-4af3-b381-f33b3af41bfd {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1056.663202] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024755, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.715646] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f047981-a705-493c-9035-627f870a4c3a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.726748] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067afade-854d-4604-90e7-520c6b323b47 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.770916] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9f0cd6-a0e4-46a4-9a52-6d30e7fbf971 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.784249] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac90c48-b284-4b18-95b0-87c9dbac9c6d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.789202] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024751, 'name': CloneVM_Task} progress is 95%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.800198] env[63372]: DEBUG nova.compute.provider_tree [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.976069] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7eab8a-d94e-4b52-818f-7c87e64f7ce9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.995473] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d4c178-f663-4374-a2d6-7d1f4fc56ba4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.005821] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024749, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.38855} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.019606] env[63372]: INFO nova.virt.vmwareapi.ds_util [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_5d8bde80-5520-473a-b3b0-c87445e752ad/OSTACK_IMG_5d8bde80-5520-473a-b3b0-c87445e752ad.vmdk to [datastore2] devstack-image-cache_base/003d2e97-266b-4459-abd7-6ca82ed4825e/003d2e97-266b-4459-abd7-6ca82ed4825e.vmdk. [ 1057.019812] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Cleaning up location [datastore2] OSTACK_IMG_5d8bde80-5520-473a-b3b0-c87445e752ad {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1057.019982] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_5d8bde80-5520-473a-b3b0-c87445e752ad {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1057.020297] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating instance '83db51f8-2c4b-4ef1-afcb-ecac63bcab9d' progress to 67 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1057.023577] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aa266df2-85f8-48f7-aa1c-abe231f92b2c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.028995] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1057.028995] env[63372]: value = "task-1024756" [ 1057.028995] env[63372]: _type = "Task" [ 1057.028995] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.036297] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024756, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.062341] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "refresh_cache-1570dee9-33d5-4af9-afe8-37cb4c331113" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.116407] env[63372]: DEBUG nova.network.neutron [req-3b37b469-fb81-44ae-bc09-36672b535815 req-4ab7b04d-138b-4214-8f43-d7b92ad80e6f service nova] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1057.162823] env[63372]: DEBUG oslo_vmware.api [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024755, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.785446} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.163086] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1057.163277] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1057.163454] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1057.163624] env[63372]: INFO nova.compute.manager [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Took 3.87 seconds to destroy the instance on the hypervisor. [ 1057.163878] env[63372]: DEBUG oslo.service.loopingcall [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.164192] env[63372]: DEBUG nova.compute.manager [-] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1057.164340] env[63372]: DEBUG nova.network.neutron [-] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1057.213037] env[63372]: DEBUG nova.network.neutron [req-3b37b469-fb81-44ae-bc09-36672b535815 req-4ab7b04d-138b-4214-8f43-d7b92ad80e6f service nova] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.273151] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024751, 'name': CloneVM_Task, 'duration_secs': 1.677466} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.273417] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Created linked-clone VM from snapshot [ 1057.274161] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d94aae2-c17c-41d9-8a8e-6abafc806bc1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.281031] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Uploading image 6735c751-59a4-497b-9de2-3765dd7aa055 {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1057.303159] env[63372]: DEBUG oslo_vmware.rw_handles [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1057.303159] env[63372]: value = "vm-227455" [ 1057.303159] env[63372]: _type = "VirtualMachine" [ 1057.303159] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1057.303470] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c2485c13-5658-4212-83d5-55664e0b2d3b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.305553] env[63372]: DEBUG nova.scheduler.client.report [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1057.313559] env[63372]: DEBUG oslo_vmware.rw_handles [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lease: (returnval){ [ 1057.313559] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523c0c7c-64dd-fdf7-cde9-88e73429f423" [ 1057.313559] env[63372]: _type = "HttpNfcLease" [ 1057.313559] env[63372]: } obtained for exporting VM: (result){ [ 1057.313559] env[63372]: value = "vm-227455" [ 1057.313559] env[63372]: _type = "VirtualMachine" [ 1057.313559] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1057.313794] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the lease: (returnval){ [ 1057.313794] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523c0c7c-64dd-fdf7-cde9-88e73429f423" [ 1057.313794] env[63372]: _type = "HttpNfcLease" [ 1057.313794] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1057.320338] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1057.320338] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523c0c7c-64dd-fdf7-cde9-88e73429f423" [ 1057.320338] env[63372]: _type = "HttpNfcLease" [ 1057.320338] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1057.540095] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024756, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.035111} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.540095] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1057.540095] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/003d2e97-266b-4459-abd7-6ca82ed4825e/003d2e97-266b-4459-abd7-6ca82ed4825e.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.540095] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/003d2e97-266b-4459-abd7-6ca82ed4825e/003d2e97-266b-4459-abd7-6ca82ed4825e.vmdk to [datastore2] 91985614-b959-401e-bb06-d67b230ee026/91985614-b959-401e-bb06-d67b230ee026.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1057.540904] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5e7b7bc-8a3c-4ec2-b65f-9426a895ac7e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.547236] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1057.547236] env[63372]: value = "task-1024758" [ 1057.547236] env[63372]: _type = "Task" [ 1057.547236] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.554355] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024758, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.579219] env[63372]: DEBUG nova.network.neutron [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Port ebc3af7a-29a8-44e6-85e7-4ae1bafa1844 binding to destination host cpu-1 is already ACTIVE {{(pid=63372) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1057.716028] env[63372]: DEBUG oslo_concurrency.lockutils [req-3b37b469-fb81-44ae-bc09-36672b535815 req-4ab7b04d-138b-4214-8f43-d7b92ad80e6f service nova] Releasing lock "refresh_cache-1570dee9-33d5-4af9-afe8-37cb4c331113" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.716422] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "refresh_cache-1570dee9-33d5-4af9-afe8-37cb4c331113" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.717029] env[63372]: DEBUG nova.network.neutron [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1057.811025] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.288s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.813021] env[63372]: INFO nova.compute.manager [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Migrating [ 1057.818492] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.073s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.818753] env[63372]: DEBUG nova.objects.instance [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lazy-loading 'resources' on Instance uuid b455b16b-3332-43bf-a91e-a350287ba5f3 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.844556] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1057.844556] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523c0c7c-64dd-fdf7-cde9-88e73429f423" [ 1057.844556] env[63372]: _type = "HttpNfcLease" [ 1057.844556] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1057.847024] env[63372]: DEBUG oslo_vmware.rw_handles [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1057.847024] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523c0c7c-64dd-fdf7-cde9-88e73429f423" [ 1057.847024] env[63372]: _type = "HttpNfcLease" [ 1057.847024] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1057.847024] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-683bc16a-2795-4559-bf40-508850b531ff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.855453] env[63372]: DEBUG oslo_vmware.rw_handles [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52029758-44a2-cfcf-7ce1-2f91ea9167c6/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1057.855654] env[63372]: DEBUG oslo_vmware.rw_handles [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52029758-44a2-cfcf-7ce1-2f91ea9167c6/disk-0.vmdk for reading. {{(pid=63372) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1057.978319] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f4ad4e26-c243-452c-9458-525db14e1954 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.061281] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024758, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.089420] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0767c5-e4ff-4398-ae98-86734e669f5f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.098068] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d34c13a3-0b30-4cb2-8d92-1613e64fd5d1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.135954] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9385d8be-e4ec-45f2-a7a9-ea683cc569d0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.146480] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ea71bb-75b1-4820-97f7-7d8de3b15ae0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.163528] env[63372]: DEBUG nova.compute.provider_tree [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.181193] env[63372]: DEBUG nova.network.neutron [-] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.251550] env[63372]: DEBUG nova.network.neutron [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1058.333664] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.333799] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.334683] env[63372]: DEBUG nova.network.neutron [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1058.409453] env[63372]: DEBUG nova.network.neutron [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Updating instance_info_cache with network_info: [{"id": "6cfbc521-36cc-4af3-b381-f33b3af41bfd", "address": "fa:16:3e:8a:3d:62", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cfbc521-36", "ovs_interfaceid": "6cfbc521-36cc-4af3-b381-f33b3af41bfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.560615] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024758, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.604871] env[63372]: DEBUG oslo_concurrency.lockutils [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.605236] env[63372]: DEBUG oslo_concurrency.lockutils [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.605451] env[63372]: DEBUG oslo_concurrency.lockutils [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.666732] env[63372]: DEBUG nova.scheduler.client.report [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1058.685611] env[63372]: INFO nova.compute.manager [-] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Took 1.52 seconds to deallocate network for instance. [ 1058.722875] env[63372]: DEBUG nova.compute.manager [req-5752b92c-222a-42ff-9e73-4ec321323d85 req-44dd788c-04ff-4e2d-aa36-c4e246d6e755 service nova] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Received event network-vif-deleted-128a0686-7ca8-427f-9cf7-c63a2f370e1f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.914584] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "refresh_cache-1570dee9-33d5-4af9-afe8-37cb4c331113" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.915227] env[63372]: DEBUG nova.compute.manager [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Instance network_info: |[{"id": "6cfbc521-36cc-4af3-b381-f33b3af41bfd", "address": "fa:16:3e:8a:3d:62", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cfbc521-36", "ovs_interfaceid": "6cfbc521-36cc-4af3-b381-f33b3af41bfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1058.916042] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:3d:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b21ab10-d886-4453-9472-9e11fb3c450d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6cfbc521-36cc-4af3-b381-f33b3af41bfd', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1058.924173] env[63372]: DEBUG oslo.service.loopingcall [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.924456] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1058.924961] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1cf4b1f2-244a-4b11-a611-8906c96da2ec {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.947174] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1058.947174] env[63372]: value = "task-1024759" [ 1058.947174] env[63372]: _type = "Task" [ 1058.947174] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.962831] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024759, 'name': CreateVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.062121] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024758, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.092472] env[63372]: DEBUG nova.network.neutron [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance_info_cache with network_info: [{"id": "7cd17ebc-3698-499b-8a5a-0507f100485c", "address": "fa:16:3e:59:5c:0e", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cd17ebc-36", "ovs_interfaceid": "7cd17ebc-3698-499b-8a5a-0507f100485c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.172692] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.354s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.201098] env[63372]: INFO nova.scheduler.client.report [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Deleted allocations for instance b455b16b-3332-43bf-a91e-a350287ba5f3 [ 1059.248146] env[63372]: INFO nova.compute.manager [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Took 0.56 seconds to detach 1 volumes for instance. [ 1059.464079] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024759, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.562332] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024758, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.595704] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.647503] env[63372]: DEBUG oslo_concurrency.lockutils [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.647837] env[63372]: DEBUG oslo_concurrency.lockutils [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.648088] env[63372]: DEBUG nova.network.neutron [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1059.711596] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a3536133-4f1e-491f-9854-3b995c2693ac tempest-AttachInterfacesTestJSON-1116050093 tempest-AttachInterfacesTestJSON-1116050093-project-member] Lock "b455b16b-3332-43bf-a91e-a350287ba5f3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.108s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.756323] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.756771] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.757049] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.780172] env[63372]: INFO nova.scheduler.client.report [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleted allocations for instance b7732621-7e58-40a3-b723-5c66df6f74a1 [ 1059.960760] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024759, 'name': CreateVM_Task, 'duration_secs': 0.751839} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.961029] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1059.961864] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.962092] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.962527] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1059.962784] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7c5256e-b2db-4a53-b551-744847849085 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.967281] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1059.967281] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52adbe4b-7638-77f7-6bc2-022910b36f14" [ 1059.967281] env[63372]: _type = "Task" [ 1059.967281] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.974970] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52adbe4b-7638-77f7-6bc2-022910b36f14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.062024] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024758, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.349402} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.062024] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/003d2e97-266b-4459-abd7-6ca82ed4825e/003d2e97-266b-4459-abd7-6ca82ed4825e.vmdk to [datastore2] 91985614-b959-401e-bb06-d67b230ee026/91985614-b959-401e-bb06-d67b230ee026.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1060.063078] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80cce1bb-4149-4e82-8ac1-d652ddebd5ca {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.087397] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 91985614-b959-401e-bb06-d67b230ee026/91985614-b959-401e-bb06-d67b230ee026.vmdk or device None with type streamOptimized {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1060.087811] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1eef06dc-636a-4801-a1ac-6492e30f3d94 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.113598] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1060.113598] env[63372]: value = "task-1024760" [ 1060.113598] env[63372]: _type = "Task" [ 1060.113598] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.122884] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024760, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.289238] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a1f63dd9-9780-439d-8276-1f7bdaf4ed19 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "b7732621-7e58-40a3-b723-5c66df6f74a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.003s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.376663] env[63372]: DEBUG nova.network.neutron [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating instance_info_cache with network_info: [{"id": "ebc3af7a-29a8-44e6-85e7-4ae1bafa1844", "address": "fa:16:3e:05:a3:85", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebc3af7a-29", "ovs_interfaceid": "ebc3af7a-29a8-44e6-85e7-4ae1bafa1844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.479376] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52adbe4b-7638-77f7-6bc2-022910b36f14, 'name': SearchDatastore_Task, 'duration_secs': 0.028046} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.479620] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.480134] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1060.480850] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.481022] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.481372] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1060.481667] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19910452-d6d0-47cb-8415-830beab3ce18 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.491741] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1060.492295] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1060.493131] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b321f6b6-1c93-46bf-bcbe-158ab4060efc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.498649] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1060.498649] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]520f6fb3-ea33-9609-914e-1b48cce623fa" [ 1060.498649] env[63372]: _type = "Task" [ 1060.498649] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.506712] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520f6fb3-ea33-9609-914e-1b48cce623fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.625627] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024760, 'name': ReconfigVM_Task, 'duration_secs': 0.288965} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.626019] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 91985614-b959-401e-bb06-d67b230ee026/91985614-b959-401e-bb06-d67b230ee026.vmdk or device None with type streamOptimized {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1060.627515] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8f18968b-b549-4f57-bf94-3b8fe848efb5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.634590] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1060.634590] env[63372]: value = "task-1024761" [ 1060.634590] env[63372]: _type = "Task" [ 1060.634590] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.645728] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024761, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.881471] env[63372]: DEBUG oslo_concurrency.lockutils [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.009452] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]520f6fb3-ea33-9609-914e-1b48cce623fa, 'name': SearchDatastore_Task, 'duration_secs': 0.010275} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.010257] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-661a74a6-957c-4cb6-b1c0-93701d63b16a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.016038] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1061.016038] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b22355-925d-a11c-420c-a8a575acf0ba" [ 1061.016038] env[63372]: _type = "Task" [ 1061.016038] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.023967] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b22355-925d-a11c-420c-a8a575acf0ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.116679] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d36d037-6ce8-4020-b4f2-bd2f6f1d8080 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.136635] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance '4a2e53e0-b68c-42be-afff-1eead017679d' progress to 0 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1061.150172] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024761, 'name': Rename_Task, 'duration_secs': 0.151455} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.150437] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1061.150657] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-28390de9-3183-45dc-86b6-8d2e04638d00 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.157544] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1061.157544] env[63372]: value = "task-1024762" [ 1061.157544] env[63372]: _type = "Task" [ 1061.157544] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.166429] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024762, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.194953] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1061.195337] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1061.195421] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Starting heal instance info cache {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1061.262977] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.263253] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.411603] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54312e4b-885e-453e-b48d-2bdf54e2c8cf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.433779] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-070b7408-44ce-4033-9e0e-e964236d1ed6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.442027] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating instance '83db51f8-2c4b-4ef1-afcb-ecac63bcab9d' progress to 83 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1061.526010] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b22355-925d-a11c-420c-a8a575acf0ba, 'name': SearchDatastore_Task, 'duration_secs': 0.034529} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.526303] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.526568] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 1570dee9-33d5-4af9-afe8-37cb4c331113/1570dee9-33d5-4af9-afe8-37cb4c331113.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1061.527258] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce50e100-0d3c-4379-9fa9-112326138180 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.534413] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1061.534413] env[63372]: value = "task-1024763" [ 1061.534413] env[63372]: _type = "Task" [ 1061.534413] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.542750] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024763, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.647049] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1061.647646] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b4820fc-86e6-41dc-8be9-bd78ee3e24fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.654904] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1061.654904] env[63372]: value = "task-1024764" [ 1061.654904] env[63372]: _type = "Task" [ 1061.654904] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.666991] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024764, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.670624] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024762, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.766086] env[63372]: DEBUG nova.compute.manager [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1061.948518] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1061.948890] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-933137ee-0bc0-4a21-8ec6-30a28f8ab521 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.958198] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1061.958198] env[63372]: value = "task-1024765" [ 1061.958198] env[63372]: _type = "Task" [ 1061.958198] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.967805] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024765, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.047495] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024763, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.167747] env[63372]: DEBUG oslo_vmware.api [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024762, 'name': PowerOnVM_Task, 'duration_secs': 0.664032} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.171404] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1062.173558] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024764, 'name': PowerOffVM_Task, 'duration_secs': 0.239215} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.174081] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1062.174290] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance '4a2e53e0-b68c-42be-afff-1eead017679d' progress to 17 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1062.300515] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.300979] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.303155] env[63372]: INFO nova.compute.claims [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1062.307228] env[63372]: DEBUG nova.compute.manager [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1062.308267] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ffd55aa-0f3c-449a-8d50-9106311e2413 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.468546] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024765, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.545184] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024763, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532675} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.545472] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 1570dee9-33d5-4af9-afe8-37cb4c331113/1570dee9-33d5-4af9-afe8-37cb4c331113.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1062.545693] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1062.545956] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea76fd0c-cd39-46fd-acc2-b6125afd28a0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.557934] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1062.557934] env[63372]: value = "task-1024766" [ 1062.557934] env[63372]: _type = "Task" [ 1062.557934] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.568462] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024766, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.681457] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1062.681764] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1062.681912] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1062.682214] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1062.682407] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1062.682662] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1062.683806] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1062.684056] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1062.684253] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1062.684793] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1062.684793] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1062.692030] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ae09423-fe22-4c74-a4fd-080e75835a63 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.704215] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Didn't find any instances for network info cache update. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1062.704468] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.704676] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.706403] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.706459] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1062.706459] env[63372]: value = "task-1024767" [ 1062.706459] env[63372]: _type = "Task" [ 1062.706459] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.706745] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.706848] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.710817] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.710982] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63372) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1062.711176] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager.update_available_resource {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.719994] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024767, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.830484] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7eda3be3-a7ea-49d7-92f2-405a18389d9a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "91985614-b959-401e-bb06-d67b230ee026" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 30.432s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.969163] env[63372]: DEBUG oslo_vmware.api [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024765, 'name': PowerOnVM_Task, 'duration_secs': 0.512326} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.969499] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1062.969704] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-731abd17-6317-4a83-b577-00a89672afc5 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating instance '83db51f8-2c4b-4ef1-afcb-ecac63bcab9d' progress to 100 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1063.068036] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024766, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079309} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.068312] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1063.069241] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853a83a9-0179-4c83-8a05-4ebc0b68ed89 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.091524] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 1570dee9-33d5-4af9-afe8-37cb4c331113/1570dee9-33d5-4af9-afe8-37cb4c331113.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.092247] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef7d6e90-c5d1-4b52-8a3b-ec6e052b1977 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.114835] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1063.114835] env[63372]: value = "task-1024768" [ 1063.114835] env[63372]: _type = "Task" [ 1063.114835] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.124150] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024768, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.217514] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.217514] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024767, 'name': ReconfigVM_Task, 'duration_secs': 0.419336} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.217514] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance '4a2e53e0-b68c-42be-afff-1eead017679d' progress to 33 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1063.494912] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c947d385-0187-441e-bed5-9218427292dd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.503932] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef89301-e116-48bc-bd67-018b21893283 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.537552] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b072c485-5cf3-4003-94c2-ca1a9b917281 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.544963] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c2baaf-fadf-418b-95a7-a906279bcf6e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.560741] env[63372]: DEBUG nova.compute.provider_tree [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1063.626524] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024768, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.724469] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1063.724789] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1063.724789] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1063.724956] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1063.725106] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1063.725347] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1063.725570] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1063.725733] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1063.725902] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1063.726082] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1063.726260] env[63372]: DEBUG nova.virt.hardware [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1063.731978] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1063.732319] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8a66cad-7672-4c7e-867b-03147baa2be1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.753308] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1063.753308] env[63372]: value = "task-1024769" [ 1063.753308] env[63372]: _type = "Task" [ 1063.753308] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.760850] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024769, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.063729] env[63372]: DEBUG nova.scheduler.client.report [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1064.125747] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024768, 'name': ReconfigVM_Task, 'duration_secs': 0.555395} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.126094] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 1570dee9-33d5-4af9-afe8-37cb4c331113/1570dee9-33d5-4af9-afe8-37cb4c331113.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.126705] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5322726c-576d-4813-8083-29a62fbf27e3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.132877] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1064.132877] env[63372]: value = "task-1024770" [ 1064.132877] env[63372]: _type = "Task" [ 1064.132877] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.141906] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024770, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.263942] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024769, 'name': ReconfigVM_Task, 'duration_secs': 0.229745} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.264546] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1064.265746] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97050e53-5418-447d-be1a-db33c58ccf64 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.296298] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 4a2e53e0-b68c-42be-afff-1eead017679d/4a2e53e0-b68c-42be-afff-1eead017679d.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1064.296650] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb0b5973-7cd2-48b9-a7ae-c283937e65d9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.318400] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1064.318400] env[63372]: value = "task-1024771" [ 1064.318400] env[63372]: _type = "Task" [ 1064.318400] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.328234] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024771, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.569567] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.268s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.570473] env[63372]: DEBUG nova.compute.manager [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1064.574539] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.358s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.574756] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.575036] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63372) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1064.577553] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-095abdd8-67f3-4417-93ad-c9ee8a846220 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.586125] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e22cfd3-2c17-4c7b-a4b5-6e12a6aab8fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.603597] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad680f9f-6c2f-46ff-a8eb-e9abec488377 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.612767] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe2aef9-8213-4eb0-8a82-62493892a8da {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.645079] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180551MB free_disk=185GB free_vcpus=48 pci_devices=None {{(pid=63372) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1064.645079] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.645079] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.664616] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024770, 'name': Rename_Task, 'duration_secs': 0.202441} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.664874] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1064.665360] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-147831f0-ec2c-4602-b29d-e8ed553c690e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.672800] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1064.672800] env[63372]: value = "task-1024772" [ 1064.672800] env[63372]: _type = "Task" [ 1064.672800] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.681575] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024772, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.828760] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024771, 'name': ReconfigVM_Task, 'duration_secs': 0.374965} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.832022] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 4a2e53e0-b68c-42be-afff-1eead017679d/4a2e53e0-b68c-42be-afff-1eead017679d.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.832022] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance '4a2e53e0-b68c-42be-afff-1eead017679d' progress to 50 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1064.899046] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa54d3b-b3f6-41b8-a423-f29b16110758 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.904952] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-20b23144-6631-45c3-8ca7-8cd14e97f84a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Suspending the VM {{(pid=63372) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1064.905260] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-6c9a2dce-dd55-49fc-a155-29c3e4987e51 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.911992] env[63372]: DEBUG oslo_vmware.api [None req-20b23144-6631-45c3-8ca7-8cd14e97f84a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1064.911992] env[63372]: value = "task-1024773" [ 1064.911992] env[63372]: _type = "Task" [ 1064.911992] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.920746] env[63372]: DEBUG oslo_vmware.api [None req-20b23144-6631-45c3-8ca7-8cd14e97f84a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024773, 'name': SuspendVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.078125] env[63372]: DEBUG nova.compute.utils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1065.080952] env[63372]: DEBUG nova.compute.manager [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1065.080952] env[63372]: DEBUG nova.network.neutron [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1065.150590] env[63372]: DEBUG nova.policy [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0cc3768e4e8c47a88565a21b7f05ad02', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '98581685387a4f1499ae6ed378af982c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1065.162962] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.163240] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.163423] env[63372]: DEBUG nova.compute.manager [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Going to confirm migration 6 {{(pid=63372) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1065.182921] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024772, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.339495] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6df8333-e517-47c0-b0a3-64cfd7d42102 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.362445] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120eca9f-7c05-4557-a02e-481d4ebf9f49 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.384278] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance '4a2e53e0-b68c-42be-afff-1eead017679d' progress to 67 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1065.422121] env[63372]: DEBUG oslo_vmware.api [None req-20b23144-6631-45c3-8ca7-8cd14e97f84a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024773, 'name': SuspendVM_Task} progress is 50%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.584546] env[63372]: DEBUG nova.compute.manager [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1065.672564] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Applying migration context for instance 4a2e53e0-b68c-42be-afff-1eead017679d as it has an incoming, in-progress migration efaa8b7b-e3cc-47f7-9e4c-0c13bcff933d. Migration status is migrating {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1065.672820] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Applying migration context for instance 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d as it has an incoming, in-progress migration 0f45e8fc-73e0-4bc6-a3a7-a5ae749ae41d. Migration status is confirming {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1065.676033] env[63372]: INFO nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating resource usage from migration efaa8b7b-e3cc-47f7-9e4c-0c13bcff933d [ 1065.676033] env[63372]: INFO nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating resource usage from migration 0f45e8fc-73e0-4bc6-a3a7-a5ae749ae41d [ 1065.702458] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024772, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.710986] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance ddca7aa5-fcc8-4835-bfdb-47781335f2cd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1065.711162] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 1724ce03-c8e2-415d-a380-59ac69fbfb57 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1065.711301] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 28eb1675-b1b8-46a6-873d-5c858b716575 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1065.711419] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance a6bd1c51-e702-48e4-b9ab-8d80db1fec9e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1065.711530] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 91985614-b959-401e-bb06-d67b230ee026 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1065.711643] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Migration 0f45e8fc-73e0-4bc6-a3a7-a5ae749ae41d is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1065.711754] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1065.711864] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 1570dee9-33d5-4af9-afe8-37cb4c331113 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1065.711972] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Migration efaa8b7b-e3cc-47f7-9e4c-0c13bcff933d is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1065.712095] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 4a2e53e0-b68c-42be-afff-1eead017679d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1065.712315] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1065.712439] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1065.712575] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2752MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1065.746099] env[63372]: DEBUG nova.network.neutron [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Successfully created port: 099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1065.842474] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.842739] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.842980] env[63372]: DEBUG nova.network.neutron [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1065.843227] env[63372]: DEBUG nova.objects.instance [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lazy-loading 'info_cache' on Instance uuid 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.894674] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0721566-9638-4311-b833-97f4d1a0d471 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.904295] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7df970-3396-4556-81e8-258bd25a6d7c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.945436] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9efd2b79-64f1-4737-820b-df6f5a6d0143 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.947053] env[63372]: DEBUG nova.network.neutron [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Port 7cd17ebc-3698-499b-8a5a-0507f100485c binding to destination host cpu-1 is already ACTIVE {{(pid=63372) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1065.954933] env[63372]: DEBUG oslo_vmware.api [None req-20b23144-6631-45c3-8ca7-8cd14e97f84a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024773, 'name': SuspendVM_Task, 'duration_secs': 0.875033} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.958101] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-20b23144-6631-45c3-8ca7-8cd14e97f84a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Suspended the VM {{(pid=63372) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1065.959991] env[63372]: DEBUG nova.compute.manager [None req-20b23144-6631-45c3-8ca7-8cd14e97f84a tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.959991] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29bdf85-69e4-4a73-8973-d0051c141737 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.963049] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de70a0f0-df18-47da-a9b4-0a7363997678 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.983311] env[63372]: DEBUG nova.compute.provider_tree [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1066.193975] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024772, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.488894] env[63372]: DEBUG nova.scheduler.client.report [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1066.597187] env[63372]: DEBUG nova.compute.manager [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1066.631687] env[63372]: DEBUG nova.virt.hardware [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1066.631949] env[63372]: DEBUG nova.virt.hardware [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1066.632172] env[63372]: DEBUG nova.virt.hardware [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1066.632728] env[63372]: DEBUG nova.virt.hardware [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1066.632728] env[63372]: DEBUG nova.virt.hardware [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1066.632728] env[63372]: DEBUG nova.virt.hardware [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1066.632863] env[63372]: DEBUG nova.virt.hardware [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1066.633104] env[63372]: DEBUG nova.virt.hardware [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1066.633358] env[63372]: DEBUG nova.virt.hardware [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1066.633541] env[63372]: DEBUG nova.virt.hardware [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1066.633718] env[63372]: DEBUG nova.virt.hardware [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1066.634676] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-007b4726-40f6-4fd8-af0b-1afd3cd41e48 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.643088] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd53b573-5a22-4e56-8c9b-b0f59995cab9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.693828] env[63372]: DEBUG oslo_vmware.api [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024772, 'name': PowerOnVM_Task, 'duration_secs': 1.830367} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.694756] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1066.694756] env[63372]: INFO nova.compute.manager [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Took 11.18 seconds to spawn the instance on the hypervisor. [ 1066.694756] env[63372]: DEBUG nova.compute.manager [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1066.696626] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acce6f5-9748-408b-8fc7-3ef968aeab5d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.982455] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "4a2e53e0-b68c-42be-afff-1eead017679d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.982788] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "4a2e53e0-b68c-42be-afff-1eead017679d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.983035] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "4a2e53e0-b68c-42be-afff-1eead017679d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.996879] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1066.997105] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.352s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.077824] env[63372]: DEBUG nova.network.neutron [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating instance_info_cache with network_info: [{"id": "ebc3af7a-29a8-44e6-85e7-4ae1bafa1844", "address": "fa:16:3e:05:a3:85", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebc3af7a-29", "ovs_interfaceid": "ebc3af7a-29a8-44e6-85e7-4ae1bafa1844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.218721] env[63372]: INFO nova.compute.manager [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Took 17.68 seconds to build instance. [ 1067.320957] env[63372]: DEBUG nova.compute.manager [req-0bf5d3af-fb7f-48c2-b01a-dd0d997996d8 req-c1ded5fe-bb94-4718-9db0-f38a184753f4 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Received event network-vif-plugged-099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1067.321605] env[63372]: DEBUG oslo_concurrency.lockutils [req-0bf5d3af-fb7f-48c2-b01a-dd0d997996d8 req-c1ded5fe-bb94-4718-9db0-f38a184753f4 service nova] Acquiring lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.321920] env[63372]: DEBUG oslo_concurrency.lockutils [req-0bf5d3af-fb7f-48c2-b01a-dd0d997996d8 req-c1ded5fe-bb94-4718-9db0-f38a184753f4 service nova] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.322179] env[63372]: DEBUG oslo_concurrency.lockutils [req-0bf5d3af-fb7f-48c2-b01a-dd0d997996d8 req-c1ded5fe-bb94-4718-9db0-f38a184753f4 service nova] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.322376] env[63372]: DEBUG nova.compute.manager [req-0bf5d3af-fb7f-48c2-b01a-dd0d997996d8 req-c1ded5fe-bb94-4718-9db0-f38a184753f4 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] No waiting events found dispatching network-vif-plugged-099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1067.322542] env[63372]: WARNING nova.compute.manager [req-0bf5d3af-fb7f-48c2-b01a-dd0d997996d8 req-c1ded5fe-bb94-4718-9db0-f38a184753f4 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Received unexpected event network-vif-plugged-099091f6-c23b-48d0-9c21-76b4bcc3ad06 for instance with vm_state building and task_state spawning. [ 1067.422994] env[63372]: DEBUG nova.network.neutron [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Successfully updated port: 099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1067.491798] env[63372]: INFO nova.compute.manager [None req-941c1832-6fd3-4d6b-b670-08e72622e2e9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Resuming [ 1067.492444] env[63372]: DEBUG nova.objects.instance [None req-941c1832-6fd3-4d6b-b670-08e72622e2e9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lazy-loading 'flavor' on Instance uuid 91985614-b959-401e-bb06-d67b230ee026 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1067.552293] env[63372]: DEBUG oslo_vmware.rw_handles [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52029758-44a2-cfcf-7ce1-2f91ea9167c6/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1067.553267] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa25b791-1ce2-4c48-a60f-e309059fc4f3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.559645] env[63372]: DEBUG oslo_vmware.rw_handles [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52029758-44a2-cfcf-7ce1-2f91ea9167c6/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1067.559813] env[63372]: ERROR oslo_vmware.rw_handles [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52029758-44a2-cfcf-7ce1-2f91ea9167c6/disk-0.vmdk due to incomplete transfer. [ 1067.560338] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7a10ba6e-71b2-4d2d-9bad-232d599aba37 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.567544] env[63372]: DEBUG oslo_vmware.rw_handles [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52029758-44a2-cfcf-7ce1-2f91ea9167c6/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1067.567743] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Uploaded image 6735c751-59a4-497b-9de2-3765dd7aa055 to the Glance image server {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1067.570073] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1067.570326] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a790d161-98f7-4856-9932-4e749b34c98b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.577202] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1067.577202] env[63372]: value = "task-1024774" [ 1067.577202] env[63372]: _type = "Task" [ 1067.577202] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.582100] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "refresh_cache-83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.582372] env[63372]: DEBUG nova.objects.instance [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lazy-loading 'migration_context' on Instance uuid 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1067.590189] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024774, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.720242] env[63372]: DEBUG oslo_concurrency.lockutils [None req-19946c84-e8cc-4e71-85ca-daf6f69c24c8 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "1570dee9-33d5-4af9-afe8-37cb4c331113" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.188s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.848961] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.849517] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.927752] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.927752] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.927752] env[63372]: DEBUG nova.network.neutron [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1068.024993] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.025233] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.025393] env[63372]: DEBUG nova.network.neutron [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1068.087826] env[63372]: DEBUG nova.objects.base [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Object Instance<83db51f8-2c4b-4ef1-afcb-ecac63bcab9d> lazy-loaded attributes: info_cache,migration_context {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1068.088223] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024774, 'name': Destroy_Task, 'duration_secs': 0.3695} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.088959] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f19fb3-fc93-4f1b-9fa6-f00a0e6cd5cb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.091436] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Destroyed the VM [ 1068.091672] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1068.091901] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6eab3c56-c709-4eb8-ac37-656eba1de64c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.111367] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f9b0875-d224-4105-b610-dc5756974ed5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.113546] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1068.113546] env[63372]: value = "task-1024775" [ 1068.113546] env[63372]: _type = "Task" [ 1068.113546] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.118020] env[63372]: DEBUG oslo_vmware.api [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1068.118020] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529c1c59-aaa5-8d26-0406-35d098657af3" [ 1068.118020] env[63372]: _type = "Task" [ 1068.118020] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.123741] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024775, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.128390] env[63372]: DEBUG oslo_vmware.api [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529c1c59-aaa5-8d26-0406-35d098657af3, 'name': SearchDatastore_Task, 'duration_secs': 0.006847} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.128673] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.128853] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.252561] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478e75a4-2673-4be7-9a7d-acda1772b543 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.259901] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-aa4b320c-602e-4487-9a5b-1652a18049ef tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Suspending the VM {{(pid=63372) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1068.260235] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-bf6e5257-4943-4476-9af8-3f69cd9a992a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.266643] env[63372]: DEBUG oslo_vmware.api [None req-aa4b320c-602e-4487-9a5b-1652a18049ef tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1068.266643] env[63372]: value = "task-1024776" [ 1068.266643] env[63372]: _type = "Task" [ 1068.266643] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.275401] env[63372]: DEBUG oslo_vmware.api [None req-aa4b320c-602e-4487-9a5b-1652a18049ef tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024776, 'name': SuspendVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.352444] env[63372]: DEBUG nova.compute.manager [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1068.459721] env[63372]: DEBUG nova.network.neutron [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1068.499944] env[63372]: DEBUG oslo_concurrency.lockutils [None req-941c1832-6fd3-4d6b-b670-08e72622e2e9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.500158] env[63372]: DEBUG oslo_concurrency.lockutils [None req-941c1832-6fd3-4d6b-b670-08e72622e2e9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquired lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.500342] env[63372]: DEBUG nova.network.neutron [None req-941c1832-6fd3-4d6b-b670-08e72622e2e9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1068.625820] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024775, 'name': RemoveSnapshot_Task, 'duration_secs': 0.3508} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.626277] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1068.626616] env[63372]: DEBUG nova.compute.manager [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1068.627434] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f982142f-ed1f-4bd1-953c-f304e4ade4b9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.689756] env[63372]: DEBUG nova.network.neutron [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Updating instance_info_cache with network_info: [{"id": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "address": "fa:16:3e:2b:d8:9b", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap099091f6-c2", "ovs_interfaceid": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.777880] env[63372]: DEBUG oslo_vmware.api [None req-aa4b320c-602e-4487-9a5b-1652a18049ef tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024776, 'name': SuspendVM_Task} progress is 45%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.818367] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5d65cb-3d27-4914-bed1-083507636a1d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.825841] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeefb48d-8447-47c9-80f6-4ddc3adca711 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.857783] env[63372]: DEBUG nova.network.neutron [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance_info_cache with network_info: [{"id": "7cd17ebc-3698-499b-8a5a-0507f100485c", "address": "fa:16:3e:59:5c:0e", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cd17ebc-36", "ovs_interfaceid": "7cd17ebc-3698-499b-8a5a-0507f100485c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.859499] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-869d030c-f4f4-44a5-9489-d19ca253f704 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.871435] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4527d4b-817c-4e78-81a6-40dcaf08e9fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.876303] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.886956] env[63372]: DEBUG nova.compute.provider_tree [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.141705] env[63372]: INFO nova.compute.manager [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Shelve offloading [ 1069.143074] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1069.143289] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87f5dae6-cc65-42b1-ba43-39dc4d0bcbae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.150425] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1069.150425] env[63372]: value = "task-1024777" [ 1069.150425] env[63372]: _type = "Task" [ 1069.150425] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.158699] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024777, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.191871] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.192207] env[63372]: DEBUG nova.compute.manager [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Instance network_info: |[{"id": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "address": "fa:16:3e:2b:d8:9b", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap099091f6-c2", "ovs_interfaceid": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1069.192774] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:d8:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '099091f6-c23b-48d0-9c21-76b4bcc3ad06', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1069.200477] env[63372]: DEBUG oslo.service.loopingcall [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.201523] env[63372]: DEBUG nova.network.neutron [None req-941c1832-6fd3-4d6b-b670-08e72622e2e9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updating instance_info_cache with network_info: [{"id": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "address": "fa:16:3e:1f:74:b1", "network": {"id": "cbdcb945-f0a5-49b1-9b18-fcd8e4ca9728", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1068757189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e517915d5e7a4e3e924cc95c55170a66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57943c40-a8", "ovs_interfaceid": "57943c40-a8c1-4d1a-bb83-d257029ce77f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.202747] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1069.203176] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c0a13cd6-77d1-4e4e-a13f-97b541ebe250 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.223109] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1069.223109] env[63372]: value = "task-1024778" [ 1069.223109] env[63372]: _type = "Task" [ 1069.223109] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.231032] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024778, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.276979] env[63372]: DEBUG oslo_vmware.api [None req-aa4b320c-602e-4487-9a5b-1652a18049ef tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024776, 'name': SuspendVM_Task} progress is 45%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.346865] env[63372]: DEBUG nova.compute.manager [req-60659918-0394-4bd7-acf4-02198465f960 req-7acc085c-92cf-4680-8f14-ca9e13e498b7 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Received event network-changed-099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.347100] env[63372]: DEBUG nova.compute.manager [req-60659918-0394-4bd7-acf4-02198465f960 req-7acc085c-92cf-4680-8f14-ca9e13e498b7 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Refreshing instance network info cache due to event network-changed-099091f6-c23b-48d0-9c21-76b4bcc3ad06. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1069.347322] env[63372]: DEBUG oslo_concurrency.lockutils [req-60659918-0394-4bd7-acf4-02198465f960 req-7acc085c-92cf-4680-8f14-ca9e13e498b7 service nova] Acquiring lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.347516] env[63372]: DEBUG oslo_concurrency.lockutils [req-60659918-0394-4bd7-acf4-02198465f960 req-7acc085c-92cf-4680-8f14-ca9e13e498b7 service nova] Acquired lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.347620] env[63372]: DEBUG nova.network.neutron [req-60659918-0394-4bd7-acf4-02198465f960 req-7acc085c-92cf-4680-8f14-ca9e13e498b7 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Refreshing network info cache for port 099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1069.365054] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.389736] env[63372]: DEBUG nova.scheduler.client.report [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1069.661100] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] VM already powered off {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1069.661100] env[63372]: DEBUG nova.compute.manager [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1069.662270] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb56a31-8594-46c6-8f03-77a535d9155c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.668107] env[63372]: DEBUG oslo_concurrency.lockutils [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.668285] env[63372]: DEBUG oslo_concurrency.lockutils [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.668460] env[63372]: DEBUG nova.network.neutron [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1069.704113] env[63372]: DEBUG oslo_concurrency.lockutils [None req-941c1832-6fd3-4d6b-b670-08e72622e2e9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Releasing lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.705059] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c19e97-6e4a-4721-8d8e-ce3e02350592 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.713738] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-941c1832-6fd3-4d6b-b670-08e72622e2e9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Resuming the VM {{(pid=63372) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1069.713995] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d335890-bce3-445a-96a9-08a9c11d2e82 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.719676] env[63372]: DEBUG oslo_vmware.api [None req-941c1832-6fd3-4d6b-b670-08e72622e2e9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1069.719676] env[63372]: value = "task-1024779" [ 1069.719676] env[63372]: _type = "Task" [ 1069.719676] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.729298] env[63372]: DEBUG oslo_vmware.api [None req-941c1832-6fd3-4d6b-b670-08e72622e2e9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024779, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.733877] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024778, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.778371] env[63372]: DEBUG oslo_vmware.api [None req-aa4b320c-602e-4487-9a5b-1652a18049ef tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024776, 'name': SuspendVM_Task, 'duration_secs': 1.184525} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.778767] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-aa4b320c-602e-4487-9a5b-1652a18049ef tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Suspended the VM {{(pid=63372) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1069.779050] env[63372]: DEBUG nova.compute.manager [None req-aa4b320c-602e-4487-9a5b-1652a18049ef tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1069.779935] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93624aff-f0af-4578-af53-f12fe4737c67 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.886532] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b26e441-f9a6-4344-8518-aaf0ae7bc36f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.909104] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e68835-d4d3-4199-89c9-745a23d40652 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.918360] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance '4a2e53e0-b68c-42be-afff-1eead017679d' progress to 83 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1070.043937] env[63372]: DEBUG nova.network.neutron [req-60659918-0394-4bd7-acf4-02198465f960 req-7acc085c-92cf-4680-8f14-ca9e13e498b7 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Updated VIF entry in instance network info cache for port 099091f6-c23b-48d0-9c21-76b4bcc3ad06. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1070.044385] env[63372]: DEBUG nova.network.neutron [req-60659918-0394-4bd7-acf4-02198465f960 req-7acc085c-92cf-4680-8f14-ca9e13e498b7 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Updating instance_info_cache with network_info: [{"id": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "address": "fa:16:3e:2b:d8:9b", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap099091f6-c2", "ovs_interfaceid": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.235102] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024778, 'name': CreateVM_Task, 'duration_secs': 0.619233} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.235398] env[63372]: DEBUG oslo_vmware.api [None req-941c1832-6fd3-4d6b-b670-08e72622e2e9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024779, 'name': PowerOnVM_Task} progress is 71%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.237656] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1070.238316] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.238484] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.238800] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1070.239337] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0b473de-83fe-41f2-bf0d-80531473053e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.243778] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1070.243778] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b9139d-9ae7-07c1-f623-37ab6f075523" [ 1070.243778] env[63372]: _type = "Task" [ 1070.243778] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.251584] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b9139d-9ae7-07c1-f623-37ab6f075523, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.386259] env[63372]: DEBUG nova.network.neutron [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Updating instance_info_cache with network_info: [{"id": "845b30fa-3e7e-4324-9479-aed7aade3f91", "address": "fa:16:3e:54:d5:15", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap845b30fa-3e", "ovs_interfaceid": "845b30fa-3e7e-4324-9479-aed7aade3f91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.414036] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.285s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.416987] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.540s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.418578] env[63372]: INFO nova.compute.claims [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1070.424066] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1070.424342] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8bbb1107-b899-4331-ae00-db08b9203c8c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.431268] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1070.431268] env[63372]: value = "task-1024780" [ 1070.431268] env[63372]: _type = "Task" [ 1070.431268] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.439769] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024780, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.547584] env[63372]: DEBUG oslo_concurrency.lockutils [req-60659918-0394-4bd7-acf4-02198465f960 req-7acc085c-92cf-4680-8f14-ca9e13e498b7 service nova] Releasing lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.732603] env[63372]: DEBUG oslo_vmware.api [None req-941c1832-6fd3-4d6b-b670-08e72622e2e9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024779, 'name': PowerOnVM_Task, 'duration_secs': 0.918626} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.732890] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-941c1832-6fd3-4d6b-b670-08e72622e2e9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Resumed the VM {{(pid=63372) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1070.733161] env[63372]: DEBUG nova.compute.manager [None req-941c1832-6fd3-4d6b-b670-08e72622e2e9 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1070.733994] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d59430-7d92-47d9-9e0e-716f836bad8d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.752880] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b9139d-9ae7-07c1-f623-37ab6f075523, 'name': SearchDatastore_Task, 'duration_secs': 0.009715} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.753217] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.753463] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1070.753720] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.753870] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.754070] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1070.754339] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76151637-ecc8-4ea1-bb9a-5b318abb40cb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.762159] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "1570dee9-33d5-4af9-afe8-37cb4c331113" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.762380] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "1570dee9-33d5-4af9-afe8-37cb4c331113" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.762584] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "1570dee9-33d5-4af9-afe8-37cb4c331113-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.762770] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "1570dee9-33d5-4af9-afe8-37cb4c331113-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.762939] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "1570dee9-33d5-4af9-afe8-37cb4c331113-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.765138] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1070.765321] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1070.766282] env[63372]: INFO nova.compute.manager [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Terminating instance [ 1070.767639] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-685b7e1e-4f94-4c5a-9a3c-fe8e1b2c949f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.770509] env[63372]: DEBUG nova.compute.manager [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1070.770734] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1070.771672] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b7633d8-94e0-4da0-b18b-34e63556f131 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.778205] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1070.778205] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b93f81-3792-0cda-4c64-d9c595bc71c3" [ 1070.778205] env[63372]: _type = "Task" [ 1070.778205] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.781369] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1070.784693] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7a0961ea-1511-49d2-a5c3-d6b0208e4d1e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.792327] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b93f81-3792-0cda-4c64-d9c595bc71c3, 'name': SearchDatastore_Task, 'duration_secs': 0.011287} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.793107] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9b367b4-2d12-404c-bbe4-0d4ea22945c8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.799024] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1070.799024] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5226b0c8-bc6d-a7a6-daaf-4525a35c6125" [ 1070.799024] env[63372]: _type = "Task" [ 1070.799024] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.806921] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5226b0c8-bc6d-a7a6-daaf-4525a35c6125, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.855950] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1070.856216] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1070.856429] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleting the datastore file [datastore2] 1570dee9-33d5-4af9-afe8-37cb4c331113 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1070.856906] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-05cc7134-c96e-4d35-831e-e1ce6584ae4b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.863508] env[63372]: DEBUG oslo_vmware.api [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1070.863508] env[63372]: value = "task-1024782" [ 1070.863508] env[63372]: _type = "Task" [ 1070.863508] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.872188] env[63372]: DEBUG oslo_vmware.api [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024782, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.889135] env[63372]: DEBUG oslo_concurrency.lockutils [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.942119] env[63372]: DEBUG oslo_vmware.api [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024780, 'name': PowerOnVM_Task, 'duration_secs': 0.386852} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.945075] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1070.945329] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dee6d967-2884-4ca3-bca0-15cf092b92b6 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance '4a2e53e0-b68c-42be-afff-1eead017679d' progress to 100 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1070.987113] env[63372]: INFO nova.scheduler.client.report [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleted allocation for migration 0f45e8fc-73e0-4bc6-a3a7-a5ae749ae41d [ 1071.268654] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1071.269913] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da69a35c-33da-4cbe-bd88-a8de77af3532 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.277917] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1071.278169] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e0e0dd7-fc0e-46ba-a654-ba81f4316168 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.308834] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5226b0c8-bc6d-a7a6-daaf-4525a35c6125, 'name': SearchDatastore_Task, 'duration_secs': 0.01001} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.309097] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.309380] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b/2b907d6c-ee2c-4cd5-a77a-babfeb41b88b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1071.309620] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-95c53ca9-3fc3-4e9e-9c89-e7e681b2fb5f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.315445] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1071.315445] env[63372]: value = "task-1024784" [ 1071.315445] env[63372]: _type = "Task" [ 1071.315445] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.323090] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024784, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.369494] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1071.369752] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1071.369976] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleting the datastore file [datastore1] 28eb1675-b1b8-46a6-873d-5c858b716575 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1071.370643] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-13a27b42-027f-4773-8844-ed4f00c47338 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.377937] env[63372]: DEBUG oslo_vmware.api [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024782, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.379117] env[63372]: DEBUG nova.compute.manager [req-9d9cf27f-8de2-4e96-9b73-d41b6fbcee10 req-246631b9-bfed-4f03-b965-6243b57eefde service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Received event network-vif-unplugged-845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1071.379343] env[63372]: DEBUG oslo_concurrency.lockutils [req-9d9cf27f-8de2-4e96-9b73-d41b6fbcee10 req-246631b9-bfed-4f03-b965-6243b57eefde service nova] Acquiring lock "28eb1675-b1b8-46a6-873d-5c858b716575-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.379550] env[63372]: DEBUG oslo_concurrency.lockutils [req-9d9cf27f-8de2-4e96-9b73-d41b6fbcee10 req-246631b9-bfed-4f03-b965-6243b57eefde service nova] Lock "28eb1675-b1b8-46a6-873d-5c858b716575-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.379914] env[63372]: DEBUG oslo_concurrency.lockutils [req-9d9cf27f-8de2-4e96-9b73-d41b6fbcee10 req-246631b9-bfed-4f03-b965-6243b57eefde service nova] Lock "28eb1675-b1b8-46a6-873d-5c858b716575-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.380233] env[63372]: DEBUG nova.compute.manager [req-9d9cf27f-8de2-4e96-9b73-d41b6fbcee10 req-246631b9-bfed-4f03-b965-6243b57eefde service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] No waiting events found dispatching network-vif-unplugged-845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1071.380607] env[63372]: WARNING nova.compute.manager [req-9d9cf27f-8de2-4e96-9b73-d41b6fbcee10 req-246631b9-bfed-4f03-b965-6243b57eefde service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Received unexpected event network-vif-unplugged-845b30fa-3e7e-4324-9479-aed7aade3f91 for instance with vm_state shelved and task_state shelving_offloading. [ 1071.380826] env[63372]: DEBUG nova.compute.manager [req-9d9cf27f-8de2-4e96-9b73-d41b6fbcee10 req-246631b9-bfed-4f03-b965-6243b57eefde service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Received event network-changed-845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1071.381212] env[63372]: DEBUG nova.compute.manager [req-9d9cf27f-8de2-4e96-9b73-d41b6fbcee10 req-246631b9-bfed-4f03-b965-6243b57eefde service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Refreshing instance network info cache due to event network-changed-845b30fa-3e7e-4324-9479-aed7aade3f91. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1071.381627] env[63372]: DEBUG oslo_concurrency.lockutils [req-9d9cf27f-8de2-4e96-9b73-d41b6fbcee10 req-246631b9-bfed-4f03-b965-6243b57eefde service nova] Acquiring lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.381834] env[63372]: DEBUG oslo_concurrency.lockutils [req-9d9cf27f-8de2-4e96-9b73-d41b6fbcee10 req-246631b9-bfed-4f03-b965-6243b57eefde service nova] Acquired lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.382015] env[63372]: DEBUG nova.network.neutron [req-9d9cf27f-8de2-4e96-9b73-d41b6fbcee10 req-246631b9-bfed-4f03-b965-6243b57eefde service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Refreshing network info cache for port 845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1071.387678] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1071.387678] env[63372]: value = "task-1024785" [ 1071.387678] env[63372]: _type = "Task" [ 1071.387678] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.396241] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024785, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.492999] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c44426e2-1dbe-4ba8-adb6-e0dc596baf32 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.330s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.577045] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a11e71-5ab7-42b3-84e3-80f98a6740e2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.584594] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6401bfe4-98b4-4a84-9df0-3d1cea9179d3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.616195] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a260ca6-2b60-4b3a-963a-4f4a96f3454b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.623922] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0b9124-dda0-4f45-aa83-cc281671ba68 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.637942] env[63372]: DEBUG nova.compute.provider_tree [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.828997] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024784, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.875323] env[63372]: DEBUG oslo_vmware.api [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024782, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.677456} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.875894] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1071.876243] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1071.876891] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1071.877264] env[63372]: INFO nova.compute.manager [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1071.877695] env[63372]: DEBUG oslo.service.loopingcall [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.878064] env[63372]: DEBUG nova.compute.manager [-] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1071.878263] env[63372]: DEBUG nova.network.neutron [-] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1071.901032] env[63372]: DEBUG oslo_vmware.api [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024785, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.496306} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.901032] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1071.901032] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1071.901032] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1071.924251] env[63372]: INFO nova.scheduler.client.report [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleted allocations for instance 28eb1675-b1b8-46a6-873d-5c858b716575 [ 1072.140643] env[63372]: DEBUG nova.scheduler.client.report [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1072.328916] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024784, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.875441} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.329576] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b/2b907d6c-ee2c-4cd5-a77a-babfeb41b88b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1072.329947] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1072.330282] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed2d7c0e-59ec-41d3-8ddb-33ee2b728c39 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.339038] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1072.339038] env[63372]: value = "task-1024786" [ 1072.339038] env[63372]: _type = "Task" [ 1072.339038] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.348486] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024786, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.428124] env[63372]: DEBUG oslo_concurrency.lockutils [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.495753] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.496328] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.497138] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.497425] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.497674] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.500973] env[63372]: INFO nova.compute.manager [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Terminating instance [ 1072.503302] env[63372]: DEBUG nova.compute.manager [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1072.503587] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1072.504712] env[63372]: DEBUG nova.network.neutron [req-9d9cf27f-8de2-4e96-9b73-d41b6fbcee10 req-246631b9-bfed-4f03-b965-6243b57eefde service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Updated VIF entry in instance network info cache for port 845b30fa-3e7e-4324-9479-aed7aade3f91. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1072.505045] env[63372]: DEBUG nova.network.neutron [req-9d9cf27f-8de2-4e96-9b73-d41b6fbcee10 req-246631b9-bfed-4f03-b965-6243b57eefde service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Updating instance_info_cache with network_info: [{"id": "845b30fa-3e7e-4324-9479-aed7aade3f91", "address": "fa:16:3e:54:d5:15", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap845b30fa-3e", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.510695] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c73d350-5c6c-40ee-8b9e-f43c2588a44f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.520666] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1072.520941] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b883846-49fc-4a01-8586-b79ef9e227d9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.528622] env[63372]: DEBUG oslo_vmware.api [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1072.528622] env[63372]: value = "task-1024787" [ 1072.528622] env[63372]: _type = "Task" [ 1072.528622] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.540071] env[63372]: DEBUG oslo_vmware.api [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024787, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.646524] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.230s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.647159] env[63372]: DEBUG nova.compute.manager [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1072.649972] env[63372]: DEBUG oslo_concurrency.lockutils [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.222s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.650218] env[63372]: DEBUG nova.objects.instance [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lazy-loading 'resources' on Instance uuid 28eb1675-b1b8-46a6-873d-5c858b716575 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.850657] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024786, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088688} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.851253] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1072.851641] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77803f0b-19f3-473c-b75f-08134cab1e85 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.874399] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b/2b907d6c-ee2c-4cd5-a77a-babfeb41b88b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1072.874781] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84926dc7-37ca-4a33-84a2-10d86876bd71 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.897012] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1072.897012] env[63372]: value = "task-1024788" [ 1072.897012] env[63372]: _type = "Task" [ 1072.897012] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.905819] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024788, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.015751] env[63372]: DEBUG oslo_concurrency.lockutils [req-9d9cf27f-8de2-4e96-9b73-d41b6fbcee10 req-246631b9-bfed-4f03-b965-6243b57eefde service nova] Releasing lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.039654] env[63372]: DEBUG oslo_vmware.api [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024787, 'name': PowerOffVM_Task, 'duration_secs': 0.487566} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.039654] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1073.039815] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1073.039939] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d7df4e48-4a27-47f6-804f-c3b10086cacb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.041674] env[63372]: DEBUG nova.network.neutron [-] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.060853] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "829591bc-9686-4c7d-895d-47a3a1e617c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.061103] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "829591bc-9686-4c7d-895d-47a3a1e617c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.129478] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1073.130015] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1073.130233] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleting the datastore file [datastore1] 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1073.130932] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e54bdd1b-bbde-493f-9cd5-1ef799b16e75 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.138937] env[63372]: DEBUG oslo_vmware.api [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1073.138937] env[63372]: value = "task-1024790" [ 1073.138937] env[63372]: _type = "Task" [ 1073.138937] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.155207] env[63372]: DEBUG nova.compute.utils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1073.156928] env[63372]: DEBUG nova.objects.instance [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lazy-loading 'numa_topology' on Instance uuid 28eb1675-b1b8-46a6-873d-5c858b716575 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.157891] env[63372]: DEBUG oslo_vmware.api [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024790, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.158631] env[63372]: DEBUG nova.compute.manager [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1073.158806] env[63372]: DEBUG nova.network.neutron [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1073.208467] env[63372]: DEBUG nova.policy [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc5c59e1419843ad90b1bc33f0a55c31', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8452e71bbbd04222bb6b868017b1b6ed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1073.407552] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024788, 'name': ReconfigVM_Task, 'duration_secs': 0.297468} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.407917] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b/2b907d6c-ee2c-4cd5-a77a-babfeb41b88b.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1073.408488] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e27c9138-e7af-45e3-acfd-69ec51cc6a71 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.415351] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1073.415351] env[63372]: value = "task-1024791" [ 1073.415351] env[63372]: _type = "Task" [ 1073.415351] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.424185] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024791, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.475774] env[63372]: DEBUG nova.network.neutron [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Port 7cd17ebc-3698-499b-8a5a-0507f100485c binding to destination host cpu-1 is already ACTIVE {{(pid=63372) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1073.476268] env[63372]: DEBUG oslo_concurrency.lockutils [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.476572] env[63372]: DEBUG oslo_concurrency.lockutils [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.476909] env[63372]: DEBUG nova.network.neutron [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1073.507417] env[63372]: DEBUG nova.compute.manager [req-ccdcd30c-e485-4ab2-8cf7-7b146b3c1a1d req-c6d0ff8d-e969-4a8b-9d15-4a0e980098e0 service nova] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Received event network-vif-deleted-6cfbc521-36cc-4af3-b381-f33b3af41bfd {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1073.545590] env[63372]: INFO nova.compute.manager [-] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Took 1.67 seconds to deallocate network for instance. [ 1073.562972] env[63372]: DEBUG nova.compute.manager [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1073.589736] env[63372]: DEBUG nova.network.neutron [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Successfully created port: ba2bc8da-ac45-4f6f-a227-c425f0b2d76d {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1073.649828] env[63372]: DEBUG oslo_vmware.api [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024790, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.339149} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.650187] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1073.650305] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1073.650504] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1073.650740] env[63372]: INFO nova.compute.manager [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1073.650983] env[63372]: DEBUG oslo.service.loopingcall [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1073.651233] env[63372]: DEBUG nova.compute.manager [-] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1073.651346] env[63372]: DEBUG nova.network.neutron [-] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1073.659773] env[63372]: DEBUG nova.compute.manager [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1073.663089] env[63372]: DEBUG nova.objects.base [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Object Instance<28eb1675-b1b8-46a6-873d-5c858b716575> lazy-loaded attributes: resources,numa_topology {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1073.844977] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceef1b24-ff18-4cb9-93e3-f4e306d9004e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.853702] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df4c0332-ece3-450c-a25e-f95a279d6090 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.889211] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db4ed72-df7e-4b8c-b5ea-3f3ca536f7a3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.896753] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2daab0de-bf1f-4589-b861-242a3f70ee05 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.914422] env[63372]: DEBUG nova.compute.provider_tree [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.923933] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024791, 'name': Rename_Task, 'duration_secs': 0.159343} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.924398] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1073.927479] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4f34ddb4-0760-4ec1-a833-22b2a27bd4e7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.933018] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1073.933018] env[63372]: value = "task-1024792" [ 1073.933018] env[63372]: _type = "Task" [ 1073.933018] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.939959] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024792, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.053779] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.089412] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.379778] env[63372]: DEBUG nova.network.neutron [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance_info_cache with network_info: [{"id": "7cd17ebc-3698-499b-8a5a-0507f100485c", "address": "fa:16:3e:59:5c:0e", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cd17ebc-36", "ovs_interfaceid": "7cd17ebc-3698-499b-8a5a-0507f100485c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.420642] env[63372]: DEBUG nova.scheduler.client.report [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1074.444410] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024792, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.603094] env[63372]: DEBUG nova.network.neutron [-] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.670573] env[63372]: DEBUG nova.compute.manager [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1074.697806] env[63372]: DEBUG nova.virt.hardware [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1074.698084] env[63372]: DEBUG nova.virt.hardware [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1074.698250] env[63372]: DEBUG nova.virt.hardware [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1074.698436] env[63372]: DEBUG nova.virt.hardware [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1074.698587] env[63372]: DEBUG nova.virt.hardware [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1074.698733] env[63372]: DEBUG nova.virt.hardware [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1074.698947] env[63372]: DEBUG nova.virt.hardware [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1074.699133] env[63372]: DEBUG nova.virt.hardware [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1074.699329] env[63372]: DEBUG nova.virt.hardware [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1074.699501] env[63372]: DEBUG nova.virt.hardware [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1074.699678] env[63372]: DEBUG nova.virt.hardware [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1074.700547] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db496553-93bf-4baa-acd5-a4dcb54c616a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.708547] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f0ea2c-d631-47ad-9d9d-21ff53a056be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.883320] env[63372]: DEBUG oslo_concurrency.lockutils [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.926263] env[63372]: DEBUG oslo_concurrency.lockutils [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.276s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.929024] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.875s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.929137] env[63372]: DEBUG nova.objects.instance [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lazy-loading 'resources' on Instance uuid 1570dee9-33d5-4af9-afe8-37cb4c331113 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1074.943976] env[63372]: DEBUG oslo_vmware.api [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024792, 'name': PowerOnVM_Task, 'duration_secs': 0.652733} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.944256] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1074.944451] env[63372]: INFO nova.compute.manager [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Took 8.35 seconds to spawn the instance on the hypervisor. [ 1074.944634] env[63372]: DEBUG nova.compute.manager [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1074.945411] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7d1c68-9989-4704-8945-c7afe08e39ba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.948791] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "28eb1675-b1b8-46a6-873d-5c858b716575" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.106114] env[63372]: INFO nova.compute.manager [-] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Took 1.45 seconds to deallocate network for instance. [ 1075.296550] env[63372]: DEBUG nova.network.neutron [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Successfully updated port: ba2bc8da-ac45-4f6f-a227-c425f0b2d76d {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1075.386710] env[63372]: DEBUG nova.compute.manager [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63372) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1075.387101] env[63372]: DEBUG oslo_concurrency.lockutils [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.437897] env[63372]: DEBUG oslo_concurrency.lockutils [None req-722fb99f-5666-45ad-9d21-913a2b5fbb12 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "28eb1675-b1b8-46a6-873d-5c858b716575" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.353s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.439219] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "28eb1675-b1b8-46a6-873d-5c858b716575" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.490s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.439219] env[63372]: INFO nova.compute.manager [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Unshelving [ 1075.464027] env[63372]: INFO nova.compute.manager [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Took 13.19 seconds to build instance. [ 1075.543366] env[63372]: DEBUG nova.compute.manager [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Received event network-vif-deleted-ebc3af7a-29a8-44e6-85e7-4ae1bafa1844 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1075.543618] env[63372]: DEBUG nova.compute.manager [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Received event network-vif-plugged-ba2bc8da-ac45-4f6f-a227-c425f0b2d76d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1075.543828] env[63372]: DEBUG oslo_concurrency.lockutils [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] Acquiring lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.544010] env[63372]: DEBUG oslo_concurrency.lockutils [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.544374] env[63372]: DEBUG oslo_concurrency.lockutils [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.544558] env[63372]: DEBUG nova.compute.manager [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] No waiting events found dispatching network-vif-plugged-ba2bc8da-ac45-4f6f-a227-c425f0b2d76d {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1075.544732] env[63372]: WARNING nova.compute.manager [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Received unexpected event network-vif-plugged-ba2bc8da-ac45-4f6f-a227-c425f0b2d76d for instance with vm_state building and task_state spawning. [ 1075.544898] env[63372]: DEBUG nova.compute.manager [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Received event network-changed-ba2bc8da-ac45-4f6f-a227-c425f0b2d76d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1075.545069] env[63372]: DEBUG nova.compute.manager [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Refreshing instance network info cache due to event network-changed-ba2bc8da-ac45-4f6f-a227-c425f0b2d76d. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1075.545271] env[63372]: DEBUG oslo_concurrency.lockutils [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] Acquiring lock "refresh_cache-9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.545408] env[63372]: DEBUG oslo_concurrency.lockutils [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] Acquired lock "refresh_cache-9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.545565] env[63372]: DEBUG nova.network.neutron [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Refreshing network info cache for port ba2bc8da-ac45-4f6f-a227-c425f0b2d76d {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1075.596909] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79abf90c-5275-4bd5-aa4d-37faaf83f6ea {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.604935] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e53fb56-2f00-482d-8025-24253bc4cdb5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.637690] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.639091] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605587d9-98e1-405b-884c-db8c3a1aba22 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.646540] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8088d9ba-2dc3-4c0c-8745-b08d20b3d2e7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.659998] env[63372]: DEBUG nova.compute.provider_tree [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.798871] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "refresh_cache-9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.964990] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a10e22c6-f1d6-4eb1-823d-30212361cdb9 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.702s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.076955] env[63372]: DEBUG nova.network.neutron [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1076.156449] env[63372]: DEBUG nova.network.neutron [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.165949] env[63372]: DEBUG nova.scheduler.client.report [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1076.459105] env[63372]: DEBUG nova.compute.manager [req-77bf9650-43a2-43ce-b5ae-31c84dfe428a req-0029deea-0d72-42ec-b6f5-037184ca28d5 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Received event network-changed-099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1076.459411] env[63372]: DEBUG nova.compute.manager [req-77bf9650-43a2-43ce-b5ae-31c84dfe428a req-0029deea-0d72-42ec-b6f5-037184ca28d5 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Refreshing instance network info cache due to event network-changed-099091f6-c23b-48d0-9c21-76b4bcc3ad06. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1076.459631] env[63372]: DEBUG oslo_concurrency.lockutils [req-77bf9650-43a2-43ce-b5ae-31c84dfe428a req-0029deea-0d72-42ec-b6f5-037184ca28d5 service nova] Acquiring lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.459786] env[63372]: DEBUG oslo_concurrency.lockutils [req-77bf9650-43a2-43ce-b5ae-31c84dfe428a req-0029deea-0d72-42ec-b6f5-037184ca28d5 service nova] Acquired lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.459954] env[63372]: DEBUG nova.network.neutron [req-77bf9650-43a2-43ce-b5ae-31c84dfe428a req-0029deea-0d72-42ec-b6f5-037184ca28d5 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Refreshing network info cache for port 099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1076.464164] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.659403] env[63372]: DEBUG oslo_concurrency.lockutils [req-b6e399f3-6221-4116-8df8-1b094ea4973c req-fb2c7a62-66af-43bc-85f1-61a93c8483cd service nova] Releasing lock "refresh_cache-9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.659804] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired lock "refresh_cache-9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.659965] env[63372]: DEBUG nova.network.neutron [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1076.671392] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.742s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.673494] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.584s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.675381] env[63372]: INFO nova.compute.claims [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1076.694255] env[63372]: INFO nova.scheduler.client.report [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleted allocations for instance 1570dee9-33d5-4af9-afe8-37cb4c331113 [ 1077.190425] env[63372]: DEBUG nova.network.neutron [req-77bf9650-43a2-43ce-b5ae-31c84dfe428a req-0029deea-0d72-42ec-b6f5-037184ca28d5 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Updated VIF entry in instance network info cache for port 099091f6-c23b-48d0-9c21-76b4bcc3ad06. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1077.190815] env[63372]: DEBUG nova.network.neutron [req-77bf9650-43a2-43ce-b5ae-31c84dfe428a req-0029deea-0d72-42ec-b6f5-037184ca28d5 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Updating instance_info_cache with network_info: [{"id": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "address": "fa:16:3e:2b:d8:9b", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap099091f6-c2", "ovs_interfaceid": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.192563] env[63372]: DEBUG nova.network.neutron [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1077.206834] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b2808f67-4120-457b-a6bd-603573da91e9 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "1570dee9-33d5-4af9-afe8-37cb4c331113" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.444s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.257605] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "91985614-b959-401e-bb06-d67b230ee026" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.257858] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "91985614-b959-401e-bb06-d67b230ee026" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.258540] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "91985614-b959-401e-bb06-d67b230ee026-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.258807] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "91985614-b959-401e-bb06-d67b230ee026-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.258987] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "91985614-b959-401e-bb06-d67b230ee026-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.261468] env[63372]: INFO nova.compute.manager [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Terminating instance [ 1077.263449] env[63372]: DEBUG nova.compute.manager [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1077.263653] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1077.264500] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26780b8f-294b-409d-9042-4d73f61ac4a6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.272523] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1077.272766] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6da9c34b-da3e-411f-b2d0-e1aee2ca7e2f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.279901] env[63372]: DEBUG oslo_vmware.api [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1077.279901] env[63372]: value = "task-1024793" [ 1077.279901] env[63372]: _type = "Task" [ 1077.279901] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.288422] env[63372]: DEBUG oslo_vmware.api [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024793, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.443491] env[63372]: DEBUG nova.network.neutron [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Updating instance_info_cache with network_info: [{"id": "ba2bc8da-ac45-4f6f-a227-c425f0b2d76d", "address": "fa:16:3e:9c:a0:66", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba2bc8da-ac", "ovs_interfaceid": "ba2bc8da-ac45-4f6f-a227-c425f0b2d76d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.696344] env[63372]: DEBUG oslo_concurrency.lockutils [req-77bf9650-43a2-43ce-b5ae-31c84dfe428a req-0029deea-0d72-42ec-b6f5-037184ca28d5 service nova] Releasing lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.791891] env[63372]: DEBUG oslo_vmware.api [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024793, 'name': PowerOffVM_Task, 'duration_secs': 0.2173} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.794118] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1077.794305] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1077.794864] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10b7471b-2d48-4669-b569-8e0e7fe6f4ae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.821650] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4dd122-ac1d-4738-8db9-e414bb2bfd86 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.828847] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635dabc0-36f4-4b0d-81b3-22ed45765767 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.858679] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3412b95b-7e73-42e3-b973-7f10e2aeea63 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.862748] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1077.862981] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1077.863172] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Deleting the datastore file [datastore2] 91985614-b959-401e-bb06-d67b230ee026 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1077.865108] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb8f2395-34d0-42e7-9be1-20ba6331bd41 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.868101] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-851b677d-7478-4043-9c83-ffd30c7b1f28 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.883286] env[63372]: DEBUG nova.compute.provider_tree [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.885358] env[63372]: DEBUG oslo_vmware.api [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for the task: (returnval){ [ 1077.885358] env[63372]: value = "task-1024795" [ 1077.885358] env[63372]: _type = "Task" [ 1077.885358] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.892613] env[63372]: DEBUG oslo_vmware.api [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024795, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.946371] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Releasing lock "refresh_cache-9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.946764] env[63372]: DEBUG nova.compute.manager [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Instance network_info: |[{"id": "ba2bc8da-ac45-4f6f-a227-c425f0b2d76d", "address": "fa:16:3e:9c:a0:66", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba2bc8da-ac", "ovs_interfaceid": "ba2bc8da-ac45-4f6f-a227-c425f0b2d76d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1077.947385] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:a0:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ba2bc8da-ac45-4f6f-a227-c425f0b2d76d', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1077.957747] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Creating folder: Project (8452e71bbbd04222bb6b868017b1b6ed). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1077.958448] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-91cbf52e-8822-46c6-a20d-f783c9eb656a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.969272] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Created folder: Project (8452e71bbbd04222bb6b868017b1b6ed) in parent group-v227230. [ 1077.969480] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Creating folder: Instances. Parent ref: group-v227458. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1077.969724] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-19530d37-78a6-4b90-a42e-d62e980d8321 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.979910] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Created folder: Instances in parent group-v227458. [ 1077.980102] env[63372]: DEBUG oslo.service.loopingcall [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1077.980457] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1077.980647] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-94b79c61-dfe1-4d70-9fd9-98a44e4ca144 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.001188] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1078.001188] env[63372]: value = "task-1024798" [ 1078.001188] env[63372]: _type = "Task" [ 1078.001188] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.009500] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024798, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.060115] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "c9e5749c-a090-4ce9-b796-852438256502" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.060518] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c9e5749c-a090-4ce9-b796-852438256502" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.387609] env[63372]: DEBUG nova.scheduler.client.report [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1078.401810] env[63372]: DEBUG oslo_vmware.api [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Task: {'id': task-1024795, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.238956} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.402792] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1078.403042] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1078.403383] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1078.403526] env[63372]: INFO nova.compute.manager [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] [instance: 91985614-b959-401e-bb06-d67b230ee026] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1078.404147] env[63372]: DEBUG oslo.service.loopingcall [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1078.404335] env[63372]: DEBUG nova.compute.manager [-] [instance: 91985614-b959-401e-bb06-d67b230ee026] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1078.404439] env[63372]: DEBUG nova.network.neutron [-] [instance: 91985614-b959-401e-bb06-d67b230ee026] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1078.512239] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024798, 'name': CreateVM_Task, 'duration_secs': 0.474378} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.512422] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1078.513168] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.513364] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.513753] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1078.514054] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffdd62c0-277d-4199-92e6-181fd17cc008 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.519526] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1078.519526] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b4aaa2-d454-b5f5-f14d-1065a0f6737a" [ 1078.519526] env[63372]: _type = "Task" [ 1078.519526] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.529018] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b4aaa2-d454-b5f5-f14d-1065a0f6737a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.563860] env[63372]: DEBUG nova.compute.manager [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1078.768656] env[63372]: DEBUG nova.compute.manager [req-f28eca4f-1a54-4ce6-b54e-eac1dea105bf req-0f3459f2-4b75-4685-a459-9dd087044b6f service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Received event network-vif-deleted-57943c40-a8c1-4d1a-bb83-d257029ce77f {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1078.768900] env[63372]: INFO nova.compute.manager [req-f28eca4f-1a54-4ce6-b54e-eac1dea105bf req-0f3459f2-4b75-4685-a459-9dd087044b6f service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Neutron deleted interface 57943c40-a8c1-4d1a-bb83-d257029ce77f; detaching it from the instance and deleting it from the info cache [ 1078.769106] env[63372]: DEBUG nova.network.neutron [req-f28eca4f-1a54-4ce6-b54e-eac1dea105bf req-0f3459f2-4b75-4685-a459-9dd087044b6f service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.896400] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.223s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.896930] env[63372]: DEBUG nova.compute.manager [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1078.899481] env[63372]: DEBUG oslo_concurrency.lockutils [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 3.513s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.030453] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b4aaa2-d454-b5f5-f14d-1065a0f6737a, 'name': SearchDatastore_Task, 'duration_secs': 0.010346} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.030757] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.031013] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1079.031364] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.031523] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.031717] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1079.031975] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-82c56178-b9d7-43bd-b34b-16e5ee2bfa63 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.041073] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1079.041319] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1079.042045] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c0f536d-14dd-4caf-a9ce-be2100045ff4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.047781] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1079.047781] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5287b15c-23be-ee75-5384-92e8b3bd5e8c" [ 1079.047781] env[63372]: _type = "Task" [ 1079.047781] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.056220] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5287b15c-23be-ee75-5384-92e8b3bd5e8c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.088164] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.246445] env[63372]: DEBUG nova.network.neutron [-] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.272606] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-be21780a-65e0-43bf-a933-ecca9647e060 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.283627] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd2efce-d568-4579-aa29-e5e8495a4ef5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.313232] env[63372]: DEBUG nova.compute.manager [req-f28eca4f-1a54-4ce6-b54e-eac1dea105bf req-0f3459f2-4b75-4685-a459-9dd087044b6f service nova] [instance: 91985614-b959-401e-bb06-d67b230ee026] Detach interface failed, port_id=57943c40-a8c1-4d1a-bb83-d257029ce77f, reason: Instance 91985614-b959-401e-bb06-d67b230ee026 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1079.403241] env[63372]: DEBUG nova.compute.utils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1079.404827] env[63372]: DEBUG nova.objects.instance [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lazy-loading 'migration_context' on Instance uuid 4a2e53e0-b68c-42be-afff-1eead017679d {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1079.406124] env[63372]: DEBUG nova.compute.manager [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1079.406309] env[63372]: DEBUG nova.network.neutron [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1079.515736] env[63372]: DEBUG nova.policy [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2253a9a409d4f1ba9176485655c0988', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dac773c2536745e285181426ae34bb96', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1079.558467] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5287b15c-23be-ee75-5384-92e8b3bd5e8c, 'name': SearchDatastore_Task, 'duration_secs': 0.010043} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.559166] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bcbd977-bb87-4ecc-b777-5f720aa5ff7c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.564890] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1079.564890] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52c2d9b1-1c57-2c03-bb21-8aa17840b83a" [ 1079.564890] env[63372]: _type = "Task" [ 1079.564890] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.572739] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c2d9b1-1c57-2c03-bb21-8aa17840b83a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.750188] env[63372]: INFO nova.compute.manager [-] [instance: 91985614-b959-401e-bb06-d67b230ee026] Took 1.35 seconds to deallocate network for instance. [ 1079.909768] env[63372]: DEBUG nova.compute.manager [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1079.932383] env[63372]: DEBUG nova.network.neutron [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Successfully created port: 6127282b-345a-481a-800c-bbc68fccdc3c {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1080.078645] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c2d9b1-1c57-2c03-bb21-8aa17840b83a, 'name': SearchDatastore_Task, 'duration_secs': 0.024501} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.078935] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.079209] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e/9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1080.079487] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b6c50bc1-b8e8-4a48-acb8-be0714f3e010 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.087338] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1080.087338] env[63372]: value = "task-1024799" [ 1080.087338] env[63372]: _type = "Task" [ 1080.087338] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.088875] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35272d7-1987-47c8-b2bc-c9d61672ee04 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.101300] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc233ab9-3364-477e-b35f-b5a20edd3513 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.104374] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024799, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.132515] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8aa6483-ac35-4541-8763-dd0a0915a699 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.141035] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e62162c-338a-40b6-a5de-2030e8bc079d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.155587] env[63372]: DEBUG nova.compute.provider_tree [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1080.256695] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.599256] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024799, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.658584] env[63372]: DEBUG nova.scheduler.client.report [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1080.929223] env[63372]: DEBUG nova.compute.manager [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1080.955880] env[63372]: DEBUG nova.virt.hardware [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1080.956181] env[63372]: DEBUG nova.virt.hardware [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1080.956433] env[63372]: DEBUG nova.virt.hardware [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1080.956678] env[63372]: DEBUG nova.virt.hardware [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1080.956815] env[63372]: DEBUG nova.virt.hardware [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1080.956942] env[63372]: DEBUG nova.virt.hardware [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1080.957168] env[63372]: DEBUG nova.virt.hardware [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1080.957334] env[63372]: DEBUG nova.virt.hardware [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1080.957505] env[63372]: DEBUG nova.virt.hardware [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1080.957670] env[63372]: DEBUG nova.virt.hardware [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1080.957845] env[63372]: DEBUG nova.virt.hardware [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.958760] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26520612-bbc6-41e8-a749-10fc6fefa046 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.969027] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e8fba0-e540-466a-8365-f288125ce8f1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.099271] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024799, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.535762} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.099475] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e/9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1081.099692] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1081.099940] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-056565a4-5813-437d-8bdd-715603524fc7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.108122] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1081.108122] env[63372]: value = "task-1024800" [ 1081.108122] env[63372]: _type = "Task" [ 1081.108122] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.115373] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024800, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.619788] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024800, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091907} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.619788] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1081.621046] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a8d36f-02b4-4da8-932d-2caef0c9f452 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.642600] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e/9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1081.642871] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bfced2ce-54e2-4c69-829f-70474f824948 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.658632] env[63372]: DEBUG nova.compute.manager [req-4376af3c-dc40-453b-a069-b177cedb0b5d req-fcfca0b1-7c15-4a7f-9666-ea8dc14f0593 service nova] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Received event network-vif-plugged-6127282b-345a-481a-800c-bbc68fccdc3c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.658845] env[63372]: DEBUG oslo_concurrency.lockutils [req-4376af3c-dc40-453b-a069-b177cedb0b5d req-fcfca0b1-7c15-4a7f-9666-ea8dc14f0593 service nova] Acquiring lock "829591bc-9686-4c7d-895d-47a3a1e617c4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.659155] env[63372]: DEBUG oslo_concurrency.lockutils [req-4376af3c-dc40-453b-a069-b177cedb0b5d req-fcfca0b1-7c15-4a7f-9666-ea8dc14f0593 service nova] Lock "829591bc-9686-4c7d-895d-47a3a1e617c4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.659235] env[63372]: DEBUG oslo_concurrency.lockutils [req-4376af3c-dc40-453b-a069-b177cedb0b5d req-fcfca0b1-7c15-4a7f-9666-ea8dc14f0593 service nova] Lock "829591bc-9686-4c7d-895d-47a3a1e617c4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.659395] env[63372]: DEBUG nova.compute.manager [req-4376af3c-dc40-453b-a069-b177cedb0b5d req-fcfca0b1-7c15-4a7f-9666-ea8dc14f0593 service nova] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] No waiting events found dispatching network-vif-plugged-6127282b-345a-481a-800c-bbc68fccdc3c {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1081.659560] env[63372]: WARNING nova.compute.manager [req-4376af3c-dc40-453b-a069-b177cedb0b5d req-fcfca0b1-7c15-4a7f-9666-ea8dc14f0593 service nova] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Received unexpected event network-vif-plugged-6127282b-345a-481a-800c-bbc68fccdc3c for instance with vm_state building and task_state spawning. [ 1081.667520] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1081.667520] env[63372]: value = "task-1024801" [ 1081.667520] env[63372]: _type = "Task" [ 1081.667520] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.669010] env[63372]: DEBUG oslo_concurrency.lockutils [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.770s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.677855] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.040s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.678065] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.679857] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.216s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.680075] env[63372]: DEBUG nova.objects.instance [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lazy-loading 'pci_requests' on Instance uuid 28eb1675-b1b8-46a6-873d-5c858b716575 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.692825] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.706552] env[63372]: INFO nova.scheduler.client.report [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleted allocations for instance 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d [ 1081.754457] env[63372]: DEBUG nova.network.neutron [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Successfully updated port: 6127282b-345a-481a-800c-bbc68fccdc3c {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1082.180367] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024801, 'name': ReconfigVM_Task, 'duration_secs': 0.282994} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.180773] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e/9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1082.181369] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1004390b-0c13-4a33-bd30-55b90b8d8d8b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.189459] env[63372]: DEBUG nova.objects.instance [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lazy-loading 'numa_topology' on Instance uuid 28eb1675-b1b8-46a6-873d-5c858b716575 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.197627] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1082.197627] env[63372]: value = "task-1024802" [ 1082.197627] env[63372]: _type = "Task" [ 1082.197627] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.215963] env[63372]: DEBUG oslo_concurrency.lockutils [None req-8f558e37-517b-4fdc-8d8e-5aaee730fd45 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "83db51f8-2c4b-4ef1-afcb-ecac63bcab9d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.720s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.217824] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024802, 'name': Rename_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.258336] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "refresh_cache-829591bc-9686-4c7d-895d-47a3a1e617c4" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.258537] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "refresh_cache-829591bc-9686-4c7d-895d-47a3a1e617c4" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.258984] env[63372]: DEBUG nova.network.neutron [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1082.699043] env[63372]: INFO nova.compute.claims [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1082.714360] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024802, 'name': Rename_Task, 'duration_secs': 0.157714} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.714602] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1082.714842] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6df18b3f-a717-4529-ae78-e442d191dcad {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.721197] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1082.721197] env[63372]: value = "task-1024803" [ 1082.721197] env[63372]: _type = "Task" [ 1082.721197] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.729064] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024803, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.792497] env[63372]: DEBUG nova.network.neutron [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1082.934641] env[63372]: DEBUG nova.network.neutron [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Updating instance_info_cache with network_info: [{"id": "6127282b-345a-481a-800c-bbc68fccdc3c", "address": "fa:16:3e:69:5a:0a", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6127282b-34", "ovs_interfaceid": "6127282b-345a-481a-800c-bbc68fccdc3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.218652] env[63372]: INFO nova.compute.manager [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Swapping old allocation on dict_keys(['c1b110ca-a185-44c2-ba1c-f05b3b420add']) held by migration efaa8b7b-e3cc-47f7-9e4c-0c13bcff933d for instance [ 1083.233652] env[63372]: DEBUG oslo_vmware.api [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024803, 'name': PowerOnVM_Task, 'duration_secs': 0.487756} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.236128] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1083.236345] env[63372]: INFO nova.compute.manager [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Took 8.57 seconds to spawn the instance on the hypervisor. [ 1083.236529] env[63372]: DEBUG nova.compute.manager [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1083.237479] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7703e02e-cc95-40b3-b566-453f26df3f6e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.245928] env[63372]: DEBUG nova.scheduler.client.report [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Overwriting current allocation {'allocations': {'c1b110ca-a185-44c2-ba1c-f05b3b420add': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 158}}, 'project_id': '5e46b6acae4a4b71bc390bd513ddd2f1', 'user_id': '6d822682f06640489a48f8577acf4389', 'consumer_generation': 1} on consumer 4a2e53e0-b68c-42be-afff-1eead017679d {{(pid=63372) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1083.318688] env[63372]: DEBUG oslo_concurrency.lockutils [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.319046] env[63372]: DEBUG oslo_concurrency.lockutils [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.319479] env[63372]: DEBUG nova.network.neutron [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1083.437569] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "refresh_cache-829591bc-9686-4c7d-895d-47a3a1e617c4" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.437907] env[63372]: DEBUG nova.compute.manager [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Instance network_info: |[{"id": "6127282b-345a-481a-800c-bbc68fccdc3c", "address": "fa:16:3e:69:5a:0a", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6127282b-34", "ovs_interfaceid": "6127282b-345a-481a-800c-bbc68fccdc3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1083.438366] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:69:5a:0a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d4ef133-b6f3-41d1-add4-92a1482195cf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6127282b-345a-481a-800c-bbc68fccdc3c', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1083.446567] env[63372]: DEBUG oslo.service.loopingcall [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1083.446798] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1083.447040] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a9b78e3c-9c30-4d49-878f-d4e3fba006c0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.468769] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1083.468769] env[63372]: value = "task-1024804" [ 1083.468769] env[63372]: _type = "Task" [ 1083.468769] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.477143] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024804, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.693609] env[63372]: DEBUG nova.compute.manager [req-432954ed-e65c-49a8-bf76-5de59a3e5544 req-a1d237af-9f94-45e0-80b6-a0094b602941 service nova] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Received event network-changed-6127282b-345a-481a-800c-bbc68fccdc3c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.693943] env[63372]: DEBUG nova.compute.manager [req-432954ed-e65c-49a8-bf76-5de59a3e5544 req-a1d237af-9f94-45e0-80b6-a0094b602941 service nova] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Refreshing instance network info cache due to event network-changed-6127282b-345a-481a-800c-bbc68fccdc3c. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1083.694267] env[63372]: DEBUG oslo_concurrency.lockutils [req-432954ed-e65c-49a8-bf76-5de59a3e5544 req-a1d237af-9f94-45e0-80b6-a0094b602941 service nova] Acquiring lock "refresh_cache-829591bc-9686-4c7d-895d-47a3a1e617c4" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.694492] env[63372]: DEBUG oslo_concurrency.lockutils [req-432954ed-e65c-49a8-bf76-5de59a3e5544 req-a1d237af-9f94-45e0-80b6-a0094b602941 service nova] Acquired lock "refresh_cache-829591bc-9686-4c7d-895d-47a3a1e617c4" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.694727] env[63372]: DEBUG nova.network.neutron [req-432954ed-e65c-49a8-bf76-5de59a3e5544 req-a1d237af-9f94-45e0-80b6-a0094b602941 service nova] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Refreshing network info cache for port 6127282b-345a-481a-800c-bbc68fccdc3c {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1083.768853] env[63372]: INFO nova.compute.manager [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Took 14.91 seconds to build instance. [ 1083.847447] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad6851e-8244-4025-841f-88cdac2231da {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.856029] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb92321f-0be6-4231-b115-52700a4c53f7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.885714] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d6d41f-1f77-4912-8b47-693c5eb8ff2f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.893590] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d6c366-daf4-45c5-a8d7-c86d0ac629a2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.907030] env[63372]: DEBUG nova.compute.provider_tree [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.978974] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024804, 'name': CreateVM_Task, 'duration_secs': 0.359312} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.979114] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1083.979785] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.979958] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.980299] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1083.980827] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-016be189-d181-4bd8-a7b4-7df3f06b05cb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.986272] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1083.986272] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]524b0001-2aa9-d94c-6d40-01ff492ed67c" [ 1083.986272] env[63372]: _type = "Task" [ 1083.986272] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.996373] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524b0001-2aa9-d94c-6d40-01ff492ed67c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.062931] env[63372]: DEBUG nova.network.neutron [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance_info_cache with network_info: [{"id": "7cd17ebc-3698-499b-8a5a-0507f100485c", "address": "fa:16:3e:59:5c:0e", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cd17ebc-36", "ovs_interfaceid": "7cd17ebc-3698-499b-8a5a-0507f100485c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.270939] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c9c6c1c5-e5e6-48a4-943d-dc08852b1cdb tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.421s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.412369] env[63372]: DEBUG nova.scheduler.client.report [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1084.458778] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.459021] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.497706] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524b0001-2aa9-d94c-6d40-01ff492ed67c, 'name': SearchDatastore_Task, 'duration_secs': 0.011021} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.498010] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.498253] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1084.498738] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.498897] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.499091] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1084.499544] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c0dae1fe-2bb9-4229-a98e-31f2fc51c7c0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.509978] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1084.510403] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1084.511155] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c370a7c-1c9f-4aea-8bdc-23029413cf0b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.517593] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1084.517593] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52334113-b85b-5465-0402-3bb3417e6d38" [ 1084.517593] env[63372]: _type = "Task" [ 1084.517593] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.522179] env[63372]: DEBUG nova.network.neutron [req-432954ed-e65c-49a8-bf76-5de59a3e5544 req-a1d237af-9f94-45e0-80b6-a0094b602941 service nova] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Updated VIF entry in instance network info cache for port 6127282b-345a-481a-800c-bbc68fccdc3c. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1084.522635] env[63372]: DEBUG nova.network.neutron [req-432954ed-e65c-49a8-bf76-5de59a3e5544 req-a1d237af-9f94-45e0-80b6-a0094b602941 service nova] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Updating instance_info_cache with network_info: [{"id": "6127282b-345a-481a-800c-bbc68fccdc3c", "address": "fa:16:3e:69:5a:0a", "network": {"id": "868f090a-43ad-4092-a0ab-7cda02b39fe1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1617577800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dac773c2536745e285181426ae34bb96", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6127282b-34", "ovs_interfaceid": "6127282b-345a-481a-800c-bbc68fccdc3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.529842] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52334113-b85b-5465-0402-3bb3417e6d38, 'name': SearchDatastore_Task, 'duration_secs': 0.009144} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.530802] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13d10cfe-91c8-48ad-8c25-2da75e4fe2b0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.537209] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1084.537209] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525a7007-89c6-e1a5-1f80-33c3c5269f40" [ 1084.537209] env[63372]: _type = "Task" [ 1084.537209] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.546209] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525a7007-89c6-e1a5-1f80-33c3c5269f40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.565284] env[63372]: DEBUG oslo_concurrency.lockutils [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-4a2e53e0-b68c-42be-afff-1eead017679d" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.565886] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1084.566076] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41ea2611-16b0-48a2-a04d-c29726f2c6ff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.574174] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1084.574174] env[63372]: value = "task-1024805" [ 1084.574174] env[63372]: _type = "Task" [ 1084.574174] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.582675] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024805, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.775277] env[63372]: DEBUG nova.compute.manager [req-2fcd5b98-94b7-4a0e-b639-6cb6697201d4 req-f0fdd443-5faa-49ec-a7be-b9c3824519e2 service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Received event network-changed-ba2bc8da-ac45-4f6f-a227-c425f0b2d76d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1084.775426] env[63372]: DEBUG nova.compute.manager [req-2fcd5b98-94b7-4a0e-b639-6cb6697201d4 req-f0fdd443-5faa-49ec-a7be-b9c3824519e2 service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Refreshing instance network info cache due to event network-changed-ba2bc8da-ac45-4f6f-a227-c425f0b2d76d. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1084.775646] env[63372]: DEBUG oslo_concurrency.lockutils [req-2fcd5b98-94b7-4a0e-b639-6cb6697201d4 req-f0fdd443-5faa-49ec-a7be-b9c3824519e2 service nova] Acquiring lock "refresh_cache-9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.775796] env[63372]: DEBUG oslo_concurrency.lockutils [req-2fcd5b98-94b7-4a0e-b639-6cb6697201d4 req-f0fdd443-5faa-49ec-a7be-b9c3824519e2 service nova] Acquired lock "refresh_cache-9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.776055] env[63372]: DEBUG nova.network.neutron [req-2fcd5b98-94b7-4a0e-b639-6cb6697201d4 req-f0fdd443-5faa-49ec-a7be-b9c3824519e2 service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Refreshing network info cache for port ba2bc8da-ac45-4f6f-a227-c425f0b2d76d {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1084.917669] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.237s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.919666] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.832s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.921334] env[63372]: INFO nova.compute.claims [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1084.961777] env[63372]: INFO nova.network.neutron [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Updating port 845b30fa-3e7e-4324-9479-aed7aade3f91 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1084.964348] env[63372]: DEBUG nova.compute.utils [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1085.025453] env[63372]: DEBUG oslo_concurrency.lockutils [req-432954ed-e65c-49a8-bf76-5de59a3e5544 req-a1d237af-9f94-45e0-80b6-a0094b602941 service nova] Releasing lock "refresh_cache-829591bc-9686-4c7d-895d-47a3a1e617c4" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.048548] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525a7007-89c6-e1a5-1f80-33c3c5269f40, 'name': SearchDatastore_Task, 'duration_secs': 0.009099} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.048851] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.049478] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 829591bc-9686-4c7d-895d-47a3a1e617c4/829591bc-9686-4c7d-895d-47a3a1e617c4.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1085.049478] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ae37b60-7261-45db-bfb7-a366c4e6e407 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.056889] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1085.056889] env[63372]: value = "task-1024806" [ 1085.056889] env[63372]: _type = "Task" [ 1085.056889] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.064966] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024806, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.081946] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024805, 'name': PowerOffVM_Task, 'duration_secs': 0.181027} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.082512] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1085.082924] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1085.083165] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1085.083328] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1085.083517] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1085.083695] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1085.083858] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1085.084077] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1085.084248] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1085.084437] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1085.084651] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1085.084839] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1085.089786] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-79ace322-9c80-4df1-a498-a6fd62a6f3c7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.107538] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1085.107538] env[63372]: value = "task-1024807" [ 1085.107538] env[63372]: _type = "Task" [ 1085.107538] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.119927] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024807, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.127538] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1085.127763] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1085.127914] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Starting heal instance info cache {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1085.128052] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Rebuilding the list of instances to heal {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1085.468242] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.542376] env[63372]: DEBUG nova.network.neutron [req-2fcd5b98-94b7-4a0e-b639-6cb6697201d4 req-f0fdd443-5faa-49ec-a7be-b9c3824519e2 service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Updated VIF entry in instance network info cache for port ba2bc8da-ac45-4f6f-a227-c425f0b2d76d. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1085.542774] env[63372]: DEBUG nova.network.neutron [req-2fcd5b98-94b7-4a0e-b639-6cb6697201d4 req-f0fdd443-5faa-49ec-a7be-b9c3824519e2 service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Updating instance_info_cache with network_info: [{"id": "ba2bc8da-ac45-4f6f-a227-c425f0b2d76d", "address": "fa:16:3e:9c:a0:66", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.180", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba2bc8da-ac", "ovs_interfaceid": "ba2bc8da-ac45-4f6f-a227-c425f0b2d76d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.567481] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024806, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.456484} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.567732] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 829591bc-9686-4c7d-895d-47a3a1e617c4/829591bc-9686-4c7d-895d-47a3a1e617c4.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1085.567944] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1085.568212] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-46d1d264-364c-47ea-a12e-be9866c3ddfe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.576259] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1085.576259] env[63372]: value = "task-1024808" [ 1085.576259] env[63372]: _type = "Task" [ 1085.576259] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.584366] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024808, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.620516] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024807, 'name': ReconfigVM_Task, 'duration_secs': 0.16835} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.621768] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a62654c-5e8c-4315-b1d8-ce3921a6a4ab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.652918] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1085.653164] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: c9e5749c-a090-4ce9-b796-852438256502] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1085.657005] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1085.657361] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1085.657598] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1085.657870] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1085.658104] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1085.658331] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1085.658632] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1085.658865] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1085.659120] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1085.659363] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1085.659652] env[63372]: DEBUG nova.virt.hardware [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1085.660800] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb4fc4e2-1122-4e57-bbe2-6e01f0780b4b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.668677] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1085.668677] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52994f49-9f03-9d80-c2ae-9785b9aebd29" [ 1085.668677] env[63372]: _type = "Task" [ 1085.668677] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.677801] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52994f49-9f03-9d80-c2ae-9785b9aebd29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.679366] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.679549] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquired lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.679730] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 91985614-b959-401e-bb06-d67b230ee026] Forcefully refreshing network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1085.679915] env[63372]: DEBUG nova.objects.instance [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lazy-loading 'info_cache' on Instance uuid 91985614-b959-401e-bb06-d67b230ee026 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.045822] env[63372]: DEBUG oslo_concurrency.lockutils [req-2fcd5b98-94b7-4a0e-b639-6cb6697201d4 req-f0fdd443-5faa-49ec-a7be-b9c3824519e2 service nova] Releasing lock "refresh_cache-9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.087860] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024808, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063374} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.090573] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1086.091677] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f3fba8-2b14-4e8c-a2d1-7d988d833f53 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.116463] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 829591bc-9686-4c7d-895d-47a3a1e617c4/829591bc-9686-4c7d-895d-47a3a1e617c4.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.119339] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0734fba-d1e3-4d90-a591-6ddfdb322c6d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.140864] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1086.140864] env[63372]: value = "task-1024809" [ 1086.140864] env[63372]: _type = "Task" [ 1086.140864] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.152175] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024809, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.181423] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52994f49-9f03-9d80-c2ae-9785b9aebd29, 'name': SearchDatastore_Task, 'duration_secs': 0.008182} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.187804] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1086.190676] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16dec1c2-2fd6-457d-b380-e1b7d64402e8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.203887] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2314fbca-4a24-4df3-9d2d-8394a1db1088 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.214915] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1086.214915] env[63372]: value = "task-1024810" [ 1086.214915] env[63372]: _type = "Task" [ 1086.214915] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.220083] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e620031-7bd6-4fbd-afd6-e9512ba2620e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.225433] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 91985614-b959-401e-bb06-d67b230ee026] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1086.253198] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024810, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.257077] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d6d0c7-666e-462f-80a8-d8755108ffda {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.266572] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b202a81-d33f-4ab9-85cb-af167198df8a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.283630] env[63372]: DEBUG nova.compute.provider_tree [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.335759] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.384038] env[63372]: DEBUG nova.compute.manager [req-1448aed0-4b75-4f89-a340-3e10ff116896 req-0100816f-f533-4870-beb0-7c456025432a service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Received event network-vif-plugged-845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1086.384300] env[63372]: DEBUG oslo_concurrency.lockutils [req-1448aed0-4b75-4f89-a340-3e10ff116896 req-0100816f-f533-4870-beb0-7c456025432a service nova] Acquiring lock "28eb1675-b1b8-46a6-873d-5c858b716575-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.384521] env[63372]: DEBUG oslo_concurrency.lockutils [req-1448aed0-4b75-4f89-a340-3e10ff116896 req-0100816f-f533-4870-beb0-7c456025432a service nova] Lock "28eb1675-b1b8-46a6-873d-5c858b716575-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.384690] env[63372]: DEBUG oslo_concurrency.lockutils [req-1448aed0-4b75-4f89-a340-3e10ff116896 req-0100816f-f533-4870-beb0-7c456025432a service nova] Lock "28eb1675-b1b8-46a6-873d-5c858b716575-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.385138] env[63372]: DEBUG nova.compute.manager [req-1448aed0-4b75-4f89-a340-3e10ff116896 req-0100816f-f533-4870-beb0-7c456025432a service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] No waiting events found dispatching network-vif-plugged-845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1086.385138] env[63372]: WARNING nova.compute.manager [req-1448aed0-4b75-4f89-a340-3e10ff116896 req-0100816f-f533-4870-beb0-7c456025432a service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Received unexpected event network-vif-plugged-845b30fa-3e7e-4324-9479-aed7aade3f91 for instance with vm_state shelved_offloaded and task_state spawning. [ 1086.480257] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.480534] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.480534] env[63372]: DEBUG nova.network.neutron [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1086.539696] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.540057] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.540328] env[63372]: INFO nova.compute.manager [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Attaching volume 2b5f42b6-add6-497e-9592-e48a741bd2f0 to /dev/sdb [ 1086.570271] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc65b791-a337-49b0-a42e-3a025c04d71b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.578231] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-948b9c59-42c6-4418-8a33-3cc1e6b26a95 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.592490] env[63372]: DEBUG nova.virt.block_device [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Updating existing volume attachment record: f067c1c5-2b24-439d-a0b4-2beebc68652c {{(pid=63372) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1086.651106] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024809, 'name': ReconfigVM_Task, 'duration_secs': 0.385983} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.651221] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 829591bc-9686-4c7d-895d-47a3a1e617c4/829591bc-9686-4c7d-895d-47a3a1e617c4.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.651918] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-afa90b82-afd4-4ca0-9a31-c8194862f5e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.659569] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1086.659569] env[63372]: value = "task-1024811" [ 1086.659569] env[63372]: _type = "Task" [ 1086.659569] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.667974] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024811, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.726447] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024810, 'name': ReconfigVM_Task, 'duration_secs': 0.216742} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.726772] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1086.727627] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745cb4df-ce10-4cc4-90bb-c09a98bbf5c6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.751627] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 4a2e53e0-b68c-42be-afff-1eead017679d/4a2e53e0-b68c-42be-afff-1eead017679d.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.752373] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e6020e3-097c-43ea-8feb-9a97a4f77e32 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.771478] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1086.771478] env[63372]: value = "task-1024813" [ 1086.771478] env[63372]: _type = "Task" [ 1086.771478] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.780414] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024813, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.786526] env[63372]: DEBUG nova.scheduler.client.report [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1086.838698] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Releasing lock "refresh_cache-91985614-b959-401e-bb06-d67b230ee026" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.838920] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 91985614-b959-401e-bb06-d67b230ee026] Updated the network info_cache for instance {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1086.839145] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.839345] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.839520] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.839672] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.839820] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.839971] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.840118] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63372) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1086.840283] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager.update_available_resource {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1087.173341] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024811, 'name': Rename_Task, 'duration_secs': 0.154928} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.173341] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1087.175312] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aff0c947-aa63-4869-8a0c-15daa2178475 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.184031] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1087.184031] env[63372]: value = "task-1024816" [ 1087.184031] env[63372]: _type = "Task" [ 1087.184031] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.192712] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024816, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.257157] env[63372]: DEBUG nova.network.neutron [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Updating instance_info_cache with network_info: [{"id": "845b30fa-3e7e-4324-9479-aed7aade3f91", "address": "fa:16:3e:54:d5:15", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap845b30fa-3e", "ovs_interfaceid": "845b30fa-3e7e-4324-9479-aed7aade3f91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.281707] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024813, 'name': ReconfigVM_Task, 'duration_secs': 0.289082} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.282020] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 4a2e53e0-b68c-42be-afff-1eead017679d/4a2e53e0-b68c-42be-afff-1eead017679d.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.282863] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9374fec3-383a-4a39-817c-d7bdbf9bbf94 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.302062] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.382s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.302597] env[63372]: DEBUG nova.compute.manager [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1087.305238] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.049s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.305546] env[63372]: DEBUG nova.objects.instance [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lazy-loading 'resources' on Instance uuid 91985614-b959-401e-bb06-d67b230ee026 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.306829] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f21f32d-cb58-4089-8d9f-5f551945ec43 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.326619] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf96063-0c2d-46b0-a144-365ae30b4555 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.347449] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.348605] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90439d62-d417-41d6-adc0-aa76bb0890dd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.356883] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1087.357170] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-63361dc5-9d47-4726-8820-9960b393385c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.364116] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1087.364116] env[63372]: value = "task-1024817" [ 1087.364116] env[63372]: _type = "Task" [ 1087.364116] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.372102] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024817, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.694158] env[63372]: DEBUG oslo_vmware.api [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024816, 'name': PowerOnVM_Task, 'duration_secs': 0.442339} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.694525] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1087.694686] env[63372]: INFO nova.compute.manager [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Took 6.77 seconds to spawn the instance on the hypervisor. [ 1087.694877] env[63372]: DEBUG nova.compute.manager [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1087.695776] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1205ee0-01d8-4af2-9288-f35a312251e5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.760105] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.791703] env[63372]: DEBUG nova.virt.hardware [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='fbd0f8d88bfbd0d0bd742582bce07b4d',container_format='bare',created_at=2024-09-30T11:43:28Z,direct_url=,disk_format='vmdk',id=6735c751-59a4-497b-9de2-3765dd7aa055,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-704582058-shelved',owner='f3d866ac9602470ea9f8d8159551acc8',properties=ImageMetaProps,protected=,size=31660544,status='active',tags=,updated_at=2024-09-30T11:43:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1087.791995] env[63372]: DEBUG nova.virt.hardware [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1087.792205] env[63372]: DEBUG nova.virt.hardware [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1087.792902] env[63372]: DEBUG nova.virt.hardware [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1087.793276] env[63372]: DEBUG nova.virt.hardware [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1087.793483] env[63372]: DEBUG nova.virt.hardware [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1087.793707] env[63372]: DEBUG nova.virt.hardware [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1087.793871] env[63372]: DEBUG nova.virt.hardware [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1087.794054] env[63372]: DEBUG nova.virt.hardware [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1087.794226] env[63372]: DEBUG nova.virt.hardware [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1087.794400] env[63372]: DEBUG nova.virt.hardware [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1087.795262] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9c502b-9119-4f8d-b644-78948b865329 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.804419] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a4d6e2-a13b-4064-93a5-c04255bbd23e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.810856] env[63372]: DEBUG nova.compute.utils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1087.815194] env[63372]: DEBUG nova.compute.manager [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1087.815364] env[63372]: DEBUG nova.network.neutron [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1087.827548] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:d5:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '845b30fa-3e7e-4324-9479-aed7aade3f91', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1087.834921] env[63372]: DEBUG oslo.service.loopingcall [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1087.838679] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1087.839098] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-770f50e8-fad8-4bab-b7cf-223df4caa278 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.862514] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1087.862514] env[63372]: value = "task-1024818" [ 1087.862514] env[63372]: _type = "Task" [ 1087.862514] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.879396] env[63372]: DEBUG oslo_vmware.api [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024817, 'name': PowerOnVM_Task, 'duration_secs': 0.359765} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.879613] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024818, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.879860] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1087.914260] env[63372]: DEBUG nova.policy [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '118d2f1de2be45cc8bb48bb75525d37e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'afe39cfedf214d50be775ef736f94da9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1087.994017] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad0aa7e-ac25-4201-8c57-8efa319d23f5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.002460] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284c3033-97f4-427b-a0ae-9dee3e8f85bc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.034166] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-965fbcb3-0ed3-45c2-86d7-0356aa04bb08 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.043119] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3270d660-ae12-4713-b963-5503839eaaa7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.059746] env[63372]: DEBUG nova.compute.provider_tree [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.172614] env[63372]: DEBUG nova.network.neutron [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Successfully created port: 4a95ce9d-92de-44b7-96b7-59f4238835c7 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1088.213993] env[63372]: INFO nova.compute.manager [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Took 14.14 seconds to build instance. [ 1088.317795] env[63372]: DEBUG nova.compute.manager [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1088.373507] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024818, 'name': CreateVM_Task} progress is 25%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.432200] env[63372]: DEBUG nova.compute.manager [req-f24a38fb-bd8c-4f45-975f-dd6f84260562 req-ac447a59-dbaf-4ae8-ad4a-0af07b4f4181 service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Received event network-changed-845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1088.432200] env[63372]: DEBUG nova.compute.manager [req-f24a38fb-bd8c-4f45-975f-dd6f84260562 req-ac447a59-dbaf-4ae8-ad4a-0af07b4f4181 service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Refreshing instance network info cache due to event network-changed-845b30fa-3e7e-4324-9479-aed7aade3f91. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1088.432200] env[63372]: DEBUG oslo_concurrency.lockutils [req-f24a38fb-bd8c-4f45-975f-dd6f84260562 req-ac447a59-dbaf-4ae8-ad4a-0af07b4f4181 service nova] Acquiring lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.432200] env[63372]: DEBUG oslo_concurrency.lockutils [req-f24a38fb-bd8c-4f45-975f-dd6f84260562 req-ac447a59-dbaf-4ae8-ad4a-0af07b4f4181 service nova] Acquired lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.432968] env[63372]: DEBUG nova.network.neutron [req-f24a38fb-bd8c-4f45-975f-dd6f84260562 req-ac447a59-dbaf-4ae8-ad4a-0af07b4f4181 service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Refreshing network info cache for port 845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1088.565302] env[63372]: DEBUG nova.scheduler.client.report [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1088.715642] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc8cd66d-95e4-41ba-862c-00c0c134c552 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "829591bc-9686-4c7d-895d-47a3a1e617c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.654s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.874162] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024818, 'name': CreateVM_Task, 'duration_secs': 0.756415} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.874372] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1088.875018] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6735c751-59a4-497b-9de2-3765dd7aa055" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.875197] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6735c751-59a4-497b-9de2-3765dd7aa055" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.875567] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6735c751-59a4-497b-9de2-3765dd7aa055" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1088.875820] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdf91101-45c7-4f7d-afc6-b6243a37780f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.881359] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1088.881359] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52de528b-567f-dea0-3abf-f940a60335c7" [ 1088.881359] env[63372]: _type = "Task" [ 1088.881359] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.889493] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52de528b-567f-dea0-3abf-f940a60335c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.893041] env[63372]: INFO nova.compute.manager [None req-364fba06-d678-4f2d-9baa-b2971ffca378 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance to original state: 'active' [ 1089.068840] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.763s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.071144] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.724s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.071430] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.071839] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63372) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1089.072871] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac3f493-c185-4efe-899f-853db7e39682 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.084879] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-125c7ae8-585f-4cb7-98c8-7a82fa348606 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.103289] env[63372]: INFO nova.scheduler.client.report [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Deleted allocations for instance 91985614-b959-401e-bb06-d67b230ee026 [ 1089.104142] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae98996-9ce6-43b6-8416-fecbb315bfc2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.118784] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd56b8a-bd1e-4464-8047-99e043abba82 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.158836] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180166MB free_disk=185GB free_vcpus=48 pci_devices=None {{(pid=63372) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1089.159013] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.159236] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.207245] env[63372]: DEBUG nova.network.neutron [req-f24a38fb-bd8c-4f45-975f-dd6f84260562 req-ac447a59-dbaf-4ae8-ad4a-0af07b4f4181 service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Updated VIF entry in instance network info cache for port 845b30fa-3e7e-4324-9479-aed7aade3f91. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1089.207245] env[63372]: DEBUG nova.network.neutron [req-f24a38fb-bd8c-4f45-975f-dd6f84260562 req-ac447a59-dbaf-4ae8-ad4a-0af07b4f4181 service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Updating instance_info_cache with network_info: [{"id": "845b30fa-3e7e-4324-9479-aed7aade3f91", "address": "fa:16:3e:54:d5:15", "network": {"id": "78b0d2b2-63ff-4bef-a2ba-43f2fbc820ca", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-756690229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f3d866ac9602470ea9f8d8159551acc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap845b30fa-3e", "ovs_interfaceid": "845b30fa-3e7e-4324-9479-aed7aade3f91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.325649] env[63372]: DEBUG nova.compute.manager [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1089.362309] env[63372]: DEBUG nova.virt.hardware [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1089.362577] env[63372]: DEBUG nova.virt.hardware [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1089.362738] env[63372]: DEBUG nova.virt.hardware [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1089.362921] env[63372]: DEBUG nova.virt.hardware [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1089.363153] env[63372]: DEBUG nova.virt.hardware [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1089.363275] env[63372]: DEBUG nova.virt.hardware [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1089.363458] env[63372]: DEBUG nova.virt.hardware [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1089.363600] env[63372]: DEBUG nova.virt.hardware [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1089.363771] env[63372]: DEBUG nova.virt.hardware [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1089.363937] env[63372]: DEBUG nova.virt.hardware [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1089.364175] env[63372]: DEBUG nova.virt.hardware [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1089.364995] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4981b0e-ade7-4562-809d-2ab1d8b5ec43 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.373194] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e4deb3-5aa8-4d38-bd6a-fdc13ae492f1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.400559] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6735c751-59a4-497b-9de2-3765dd7aa055" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.401519] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Processing image 6735c751-59a4-497b-9de2-3765dd7aa055 {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1089.401519] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6735c751-59a4-497b-9de2-3765dd7aa055/6735c751-59a4-497b-9de2-3765dd7aa055.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.401519] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6735c751-59a4-497b-9de2-3765dd7aa055/6735c751-59a4-497b-9de2-3765dd7aa055.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.401519] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1089.404816] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e57321ef-b458-4263-9765-7787d871c1e3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.428302] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1089.428560] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1089.429771] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46d6d236-f655-4514-acfd-4280b15fb286 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.436165] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1089.436165] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52642fe9-7d0a-db03-9121-17ce73e32b34" [ 1089.436165] env[63372]: _type = "Task" [ 1089.436165] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.445500] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52642fe9-7d0a-db03-9121-17ce73e32b34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.613215] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dbc20fc6-38ec-41a9-9a03-35d201c2ad10 tempest-ServersNegativeTestJSON-1860392579 tempest-ServersNegativeTestJSON-1860392579-project-member] Lock "91985614-b959-401e-bb06-d67b230ee026" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.355s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.711436] env[63372]: DEBUG oslo_concurrency.lockutils [req-f24a38fb-bd8c-4f45-975f-dd6f84260562 req-ac447a59-dbaf-4ae8-ad4a-0af07b4f4181 service nova] Releasing lock "refresh_cache-28eb1675-b1b8-46a6-873d-5c858b716575" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.740445] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "829591bc-9686-4c7d-895d-47a3a1e617c4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.740814] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "829591bc-9686-4c7d-895d-47a3a1e617c4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.741067] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "829591bc-9686-4c7d-895d-47a3a1e617c4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.741234] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "829591bc-9686-4c7d-895d-47a3a1e617c4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.741501] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "829591bc-9686-4c7d-895d-47a3a1e617c4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.743951] env[63372]: INFO nova.compute.manager [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Terminating instance [ 1089.745867] env[63372]: DEBUG nova.compute.manager [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1089.746094] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1089.746998] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c01b629-3553-42cb-9565-92c060dc535d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.756118] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1089.756454] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e953737-aedf-46ba-a369-e3cc0a12d809 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.763864] env[63372]: DEBUG oslo_vmware.api [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1089.763864] env[63372]: value = "task-1024820" [ 1089.763864] env[63372]: _type = "Task" [ 1089.763864] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.773258] env[63372]: DEBUG oslo_vmware.api [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024820, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.833464] env[63372]: DEBUG nova.network.neutron [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Successfully updated port: 4a95ce9d-92de-44b7-96b7-59f4238835c7 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1089.948949] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Preparing fetch location {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1089.949331] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Fetch image to [datastore1] OSTACK_IMG_2ff065e3-d625-4a44-9d78-7cfa53184c82/OSTACK_IMG_2ff065e3-d625-4a44-9d78-7cfa53184c82.vmdk {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1089.949574] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Downloading stream optimized image 6735c751-59a4-497b-9de2-3765dd7aa055 to [datastore1] OSTACK_IMG_2ff065e3-d625-4a44-9d78-7cfa53184c82/OSTACK_IMG_2ff065e3-d625-4a44-9d78-7cfa53184c82.vmdk on the data store datastore1 as vApp {{(pid=63372) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1089.949787] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Downloading image file data 6735c751-59a4-497b-9de2-3765dd7aa055 to the ESX as VM named 'OSTACK_IMG_2ff065e3-d625-4a44-9d78-7cfa53184c82' {{(pid=63372) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1089.985881] env[63372]: DEBUG oslo_concurrency.lockutils [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "4a2e53e0-b68c-42be-afff-1eead017679d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.985881] env[63372]: DEBUG oslo_concurrency.lockutils [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "4a2e53e0-b68c-42be-afff-1eead017679d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.985881] env[63372]: DEBUG oslo_concurrency.lockutils [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "4a2e53e0-b68c-42be-afff-1eead017679d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.985881] env[63372]: DEBUG oslo_concurrency.lockutils [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "4a2e53e0-b68c-42be-afff-1eead017679d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.985881] env[63372]: DEBUG oslo_concurrency.lockutils [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "4a2e53e0-b68c-42be-afff-1eead017679d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.005343] env[63372]: INFO nova.compute.manager [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Terminating instance [ 1090.008591] env[63372]: DEBUG nova.compute.manager [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1090.008945] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1090.009990] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfeca998-b78e-45e8-9e92-36c67043d46b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.019564] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1090.021881] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f571583-db17-4d1b-8e88-9657ee199f1b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.031030] env[63372]: DEBUG oslo_vmware.api [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1090.031030] env[63372]: value = "task-1024821" [ 1090.031030] env[63372]: _type = "Task" [ 1090.031030] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.040467] env[63372]: DEBUG oslo_vmware.api [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024821, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.051068] env[63372]: DEBUG oslo_vmware.rw_handles [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1090.051068] env[63372]: value = "resgroup-9" [ 1090.051068] env[63372]: _type = "ResourcePool" [ 1090.051068] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1090.051362] env[63372]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-6c999a87-929f-4365-8ce0-5b67673699a5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.075474] env[63372]: DEBUG oslo_vmware.rw_handles [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lease: (returnval){ [ 1090.075474] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fdd27c-f750-e7b1-aa31-23ffca026a31" [ 1090.075474] env[63372]: _type = "HttpNfcLease" [ 1090.075474] env[63372]: } obtained for vApp import into resource pool (val){ [ 1090.075474] env[63372]: value = "resgroup-9" [ 1090.075474] env[63372]: _type = "ResourcePool" [ 1090.075474] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1090.075916] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the lease: (returnval){ [ 1090.075916] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fdd27c-f750-e7b1-aa31-23ffca026a31" [ 1090.075916] env[63372]: _type = "HttpNfcLease" [ 1090.075916] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1090.083231] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1090.083231] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fdd27c-f750-e7b1-aa31-23ffca026a31" [ 1090.083231] env[63372]: _type = "HttpNfcLease" [ 1090.083231] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1090.191152] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance ddca7aa5-fcc8-4835-bfdb-47781335f2cd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1090.191355] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 1724ce03-c8e2-415d-a380-59ac69fbfb57 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1090.191845] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance a6bd1c51-e702-48e4-b9ab-8d80db1fec9e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1090.191937] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 4a2e53e0-b68c-42be-afff-1eead017679d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1090.192092] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1090.192222] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1090.192368] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 829591bc-9686-4c7d-895d-47a3a1e617c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1090.192518] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 28eb1675-b1b8-46a6-873d-5c858b716575 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1090.192638] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance c9e5749c-a090-4ce9-b796-852438256502 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1090.192875] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1090.193132] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1090.277428] env[63372]: DEBUG oslo_vmware.api [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024820, 'name': PowerOffVM_Task, 'duration_secs': 0.207291} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.280958] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1090.281269] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1090.282487] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc33b769-048e-45b0-a7f1-f283f12537e7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.329213] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1638f3d5-6772-49a4-8a32-af5158dc3e99 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.337539] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.337686] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.337841] env[63372]: DEBUG nova.network.neutron [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1090.343933] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad7a6be-b4d5-4af7-9793-2ffbc98a50e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.354895] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1090.355159] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1090.355350] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleting the datastore file [datastore1] 829591bc-9686-4c7d-895d-47a3a1e617c4 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1090.355900] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-51946f67-a909-4f90-b76f-a9ceb705f65a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.384518] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebaf0111-b57a-494a-9909-5d7d07d67ce4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.390624] env[63372]: DEBUG oslo_vmware.api [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for the task: (returnval){ [ 1090.390624] env[63372]: value = "task-1024824" [ 1090.390624] env[63372]: _type = "Task" [ 1090.390624] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.397199] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530ab3eb-2064-4e49-a85d-98af96f0dbc9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.407779] env[63372]: DEBUG oslo_vmware.api [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024824, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.417816] env[63372]: DEBUG nova.compute.provider_tree [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1090.460241] env[63372]: DEBUG nova.compute.manager [req-6aa432d4-df65-4247-b5aa-71360a6997a3 req-babd6de3-17f5-439c-bbc4-e5ec4d1d215d service nova] [instance: c9e5749c-a090-4ce9-b796-852438256502] Received event network-vif-plugged-4a95ce9d-92de-44b7-96b7-59f4238835c7 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1090.460483] env[63372]: DEBUG oslo_concurrency.lockutils [req-6aa432d4-df65-4247-b5aa-71360a6997a3 req-babd6de3-17f5-439c-bbc4-e5ec4d1d215d service nova] Acquiring lock "c9e5749c-a090-4ce9-b796-852438256502-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.460724] env[63372]: DEBUG oslo_concurrency.lockutils [req-6aa432d4-df65-4247-b5aa-71360a6997a3 req-babd6de3-17f5-439c-bbc4-e5ec4d1d215d service nova] Lock "c9e5749c-a090-4ce9-b796-852438256502-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.460884] env[63372]: DEBUG oslo_concurrency.lockutils [req-6aa432d4-df65-4247-b5aa-71360a6997a3 req-babd6de3-17f5-439c-bbc4-e5ec4d1d215d service nova] Lock "c9e5749c-a090-4ce9-b796-852438256502-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.461074] env[63372]: DEBUG nova.compute.manager [req-6aa432d4-df65-4247-b5aa-71360a6997a3 req-babd6de3-17f5-439c-bbc4-e5ec4d1d215d service nova] [instance: c9e5749c-a090-4ce9-b796-852438256502] No waiting events found dispatching network-vif-plugged-4a95ce9d-92de-44b7-96b7-59f4238835c7 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1090.461252] env[63372]: WARNING nova.compute.manager [req-6aa432d4-df65-4247-b5aa-71360a6997a3 req-babd6de3-17f5-439c-bbc4-e5ec4d1d215d service nova] [instance: c9e5749c-a090-4ce9-b796-852438256502] Received unexpected event network-vif-plugged-4a95ce9d-92de-44b7-96b7-59f4238835c7 for instance with vm_state building and task_state spawning. [ 1090.461469] env[63372]: DEBUG nova.compute.manager [req-6aa432d4-df65-4247-b5aa-71360a6997a3 req-babd6de3-17f5-439c-bbc4-e5ec4d1d215d service nova] [instance: c9e5749c-a090-4ce9-b796-852438256502] Received event network-changed-4a95ce9d-92de-44b7-96b7-59f4238835c7 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1090.461659] env[63372]: DEBUG nova.compute.manager [req-6aa432d4-df65-4247-b5aa-71360a6997a3 req-babd6de3-17f5-439c-bbc4-e5ec4d1d215d service nova] [instance: c9e5749c-a090-4ce9-b796-852438256502] Refreshing instance network info cache due to event network-changed-4a95ce9d-92de-44b7-96b7-59f4238835c7. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1090.461848] env[63372]: DEBUG oslo_concurrency.lockutils [req-6aa432d4-df65-4247-b5aa-71360a6997a3 req-babd6de3-17f5-439c-bbc4-e5ec4d1d215d service nova] Acquiring lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.543070] env[63372]: DEBUG oslo_vmware.api [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024821, 'name': PowerOffVM_Task, 'duration_secs': 0.198536} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.543070] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1090.543070] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1090.543070] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-48649a1f-de99-4f85-98c0-1dcae8be8268 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.591081] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1090.591081] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fdd27c-f750-e7b1-aa31-23ffca026a31" [ 1090.591081] env[63372]: _type = "HttpNfcLease" [ 1090.591081] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1090.591467] env[63372]: DEBUG oslo_vmware.rw_handles [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1090.591467] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fdd27c-f750-e7b1-aa31-23ffca026a31" [ 1090.591467] env[63372]: _type = "HttpNfcLease" [ 1090.591467] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1090.592273] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b73b3a-7cd5-460d-a427-468dbb210561 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.604992] env[63372]: DEBUG oslo_vmware.rw_handles [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ea957-870e-6d06-f406-d1f06f33e894/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1090.605222] env[63372]: DEBUG oslo_vmware.rw_handles [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Creating HTTP connection to write to file with size = 31660544 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ea957-870e-6d06-f406-d1f06f33e894/disk-0.vmdk. {{(pid=63372) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1090.668834] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1090.669172] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1090.669369] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleting the datastore file [datastore1] 4a2e53e0-b68c-42be-afff-1eead017679d {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1090.671256] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-77fb82a0-079e-4aad-a0ee-7fc65c1ee755 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.677044] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1de8b4b9-0951-4665-9707-f91f7e564fb0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.684294] env[63372]: DEBUG oslo_vmware.api [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1090.684294] env[63372]: value = "task-1024826" [ 1090.684294] env[63372]: _type = "Task" [ 1090.684294] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.694743] env[63372]: DEBUG oslo_vmware.api [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024826, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.890572] env[63372]: DEBUG nova.network.neutron [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1090.905659] env[63372]: DEBUG oslo_vmware.api [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Task: {'id': task-1024824, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18434} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.906011] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1090.906247] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1090.906439] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1090.906657] env[63372]: INFO nova.compute.manager [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1090.906954] env[63372]: DEBUG oslo.service.loopingcall [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1090.907195] env[63372]: DEBUG nova.compute.manager [-] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1090.907290] env[63372]: DEBUG nova.network.neutron [-] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1090.921466] env[63372]: DEBUG nova.scheduler.client.report [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1091.149944] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Volume attach. Driver type: vmdk {{(pid=63372) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1091.149944] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227463', 'volume_id': '2b5f42b6-add6-497e-9592-e48a741bd2f0', 'name': 'volume-2b5f42b6-add6-497e-9592-e48a741bd2f0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a6bd1c51-e702-48e4-b9ab-8d80db1fec9e', 'attached_at': '', 'detached_at': '', 'volume_id': '2b5f42b6-add6-497e-9592-e48a741bd2f0', 'serial': '2b5f42b6-add6-497e-9592-e48a741bd2f0'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1091.149944] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fea3063-9fa7-4c45-98ad-a5b41de86574 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.176203] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34bf3241-d6e9-44c5-ab64-4645cb8435b5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.204778] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] volume-2b5f42b6-add6-497e-9592-e48a741bd2f0/volume-2b5f42b6-add6-497e-9592-e48a741bd2f0.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1091.207940] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b1ddd2e-968f-4cfd-96ad-7917dfad69d5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.231477] env[63372]: DEBUG oslo_vmware.api [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024826, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162796} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.233250] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1091.233250] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1091.233492] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1091.233737] env[63372]: INFO nova.compute.manager [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Took 1.22 seconds to destroy the instance on the hypervisor. [ 1091.234048] env[63372]: DEBUG oslo.service.loopingcall [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.234337] env[63372]: DEBUG oslo_vmware.api [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1091.234337] env[63372]: value = "task-1024827" [ 1091.234337] env[63372]: _type = "Task" [ 1091.234337] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.235265] env[63372]: DEBUG nova.compute.manager [-] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1091.235366] env[63372]: DEBUG nova.network.neutron [-] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1091.247722] env[63372]: DEBUG nova.compute.manager [req-59062188-750e-49e4-b1b1-273d7f210df6 req-85c1856e-ffba-4634-b1c6-3c56e4faac6a service nova] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Received event network-vif-deleted-6127282b-345a-481a-800c-bbc68fccdc3c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1091.247940] env[63372]: INFO nova.compute.manager [req-59062188-750e-49e4-b1b1-273d7f210df6 req-85c1856e-ffba-4634-b1c6-3c56e4faac6a service nova] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Neutron deleted interface 6127282b-345a-481a-800c-bbc68fccdc3c; detaching it from the instance and deleting it from the info cache [ 1091.248259] env[63372]: DEBUG nova.network.neutron [req-59062188-750e-49e4-b1b1-273d7f210df6 req-85c1856e-ffba-4634-b1c6-3c56e4faac6a service nova] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.261571] env[63372]: DEBUG oslo_vmware.api [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024827, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.428503] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1091.428788] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.269s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.431846] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.432021] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Cleaning up deleted instances {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1091.461481] env[63372]: DEBUG nova.network.neutron [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating instance_info_cache with network_info: [{"id": "4a95ce9d-92de-44b7-96b7-59f4238835c7", "address": "fa:16:3e:51:87:f7", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a95ce9d-92", "ovs_interfaceid": "4a95ce9d-92de-44b7-96b7-59f4238835c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.727483] env[63372]: DEBUG nova.network.neutron [-] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.752749] env[63372]: DEBUG oslo_vmware.api [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.752949] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-edf736b3-9b91-41cc-ba36-c2233cd986ed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.767250] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec337d98-fed2-4f07-93fe-5d65d7783850 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.808206] env[63372]: DEBUG nova.compute.manager [req-59062188-750e-49e4-b1b1-273d7f210df6 req-85c1856e-ffba-4634-b1c6-3c56e4faac6a service nova] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Detach interface failed, port_id=6127282b-345a-481a-800c-bbc68fccdc3c, reason: Instance 829591bc-9686-4c7d-895d-47a3a1e617c4 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1091.954496] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] There are 56 instances to clean {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1091.954761] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 1570dee9-33d5-4af9-afe8-37cb4c331113] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.963545] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.963851] env[63372]: DEBUG nova.compute.manager [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Instance network_info: |[{"id": "4a95ce9d-92de-44b7-96b7-59f4238835c7", "address": "fa:16:3e:51:87:f7", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a95ce9d-92", "ovs_interfaceid": "4a95ce9d-92de-44b7-96b7-59f4238835c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1091.964719] env[63372]: DEBUG oslo_concurrency.lockutils [req-6aa432d4-df65-4247-b5aa-71360a6997a3 req-babd6de3-17f5-439c-bbc4-e5ec4d1d215d service nova] Acquired lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.964982] env[63372]: DEBUG nova.network.neutron [req-6aa432d4-df65-4247-b5aa-71360a6997a3 req-babd6de3-17f5-439c-bbc4-e5ec4d1d215d service nova] [instance: c9e5749c-a090-4ce9-b796-852438256502] Refreshing network info cache for port 4a95ce9d-92de-44b7-96b7-59f4238835c7 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1091.966598] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:87:f7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b21ab10-d886-4453-9472-9e11fb3c450d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4a95ce9d-92de-44b7-96b7-59f4238835c7', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1091.976187] env[63372]: DEBUG oslo.service.loopingcall [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.982516] env[63372]: DEBUG oslo_vmware.rw_handles [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Completed reading data from the image iterator. {{(pid=63372) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1091.982797] env[63372]: DEBUG oslo_vmware.rw_handles [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ea957-870e-6d06-f406-d1f06f33e894/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1091.983112] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9e5749c-a090-4ce9-b796-852438256502] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1091.984314] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593188cc-efc2-40e4-91e6-0a859b3e2d18 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.988363] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56b12b19-d3cc-4c3d-a55b-cc4daa8ffbdf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.009699] env[63372]: DEBUG oslo_vmware.rw_handles [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ea957-870e-6d06-f406-d1f06f33e894/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1092.009950] env[63372]: DEBUG oslo_vmware.rw_handles [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ea957-870e-6d06-f406-d1f06f33e894/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1092.010257] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-925fcde1-8d09-4911-9834-1356ce8dc224 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.013469] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1092.013469] env[63372]: value = "task-1024828" [ 1092.013469] env[63372]: _type = "Task" [ 1092.013469] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.024722] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024828, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.234593] env[63372]: INFO nova.compute.manager [-] [instance: 829591bc-9686-4c7d-895d-47a3a1e617c4] Took 1.33 seconds to deallocate network for instance. [ 1092.239342] env[63372]: DEBUG nova.network.neutron [-] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.258102] env[63372]: DEBUG oslo_vmware.api [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024827, 'name': ReconfigVM_Task, 'duration_secs': 0.612834} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.258441] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Reconfigured VM instance instance-00000068 to attach disk [datastore1] volume-2b5f42b6-add6-497e-9592-e48a741bd2f0/volume-2b5f42b6-add6-497e-9592-e48a741bd2f0.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1092.264157] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03c573f0-b6da-4905-b567-cdbb65b5f41d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.284650] env[63372]: DEBUG oslo_vmware.api [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1092.284650] env[63372]: value = "task-1024829" [ 1092.284650] env[63372]: _type = "Task" [ 1092.284650] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.286172] env[63372]: DEBUG nova.network.neutron [req-6aa432d4-df65-4247-b5aa-71360a6997a3 req-babd6de3-17f5-439c-bbc4-e5ec4d1d215d service nova] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updated VIF entry in instance network info cache for port 4a95ce9d-92de-44b7-96b7-59f4238835c7. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1092.286514] env[63372]: DEBUG nova.network.neutron [req-6aa432d4-df65-4247-b5aa-71360a6997a3 req-babd6de3-17f5-439c-bbc4-e5ec4d1d215d service nova] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating instance_info_cache with network_info: [{"id": "4a95ce9d-92de-44b7-96b7-59f4238835c7", "address": "fa:16:3e:51:87:f7", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a95ce9d-92", "ovs_interfaceid": "4a95ce9d-92de-44b7-96b7-59f4238835c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.298413] env[63372]: DEBUG oslo_vmware.api [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024829, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.367337] env[63372]: DEBUG oslo_vmware.rw_handles [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ea957-870e-6d06-f406-d1f06f33e894/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1092.367566] env[63372]: INFO nova.virt.vmwareapi.images [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Downloaded image file data 6735c751-59a4-497b-9de2-3765dd7aa055 [ 1092.368926] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89067f17-0be7-4634-8972-e5051298a81c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.389072] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0ac68cd8-6c03-4c5d-a403-0fef8ec1f1fd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.423724] env[63372]: INFO nova.virt.vmwareapi.images [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] The imported VM was unregistered [ 1092.426543] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Caching image {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1092.427018] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Creating directory with path [datastore1] devstack-image-cache_base/6735c751-59a4-497b-9de2-3765dd7aa055 {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1092.427135] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97e85bea-6fd1-44c9-8b2c-489c308705d1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.440895] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Created directory with path [datastore1] devstack-image-cache_base/6735c751-59a4-497b-9de2-3765dd7aa055 {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1092.441180] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_2ff065e3-d625-4a44-9d78-7cfa53184c82/OSTACK_IMG_2ff065e3-d625-4a44-9d78-7cfa53184c82.vmdk to [datastore1] devstack-image-cache_base/6735c751-59a4-497b-9de2-3765dd7aa055/6735c751-59a4-497b-9de2-3765dd7aa055.vmdk. {{(pid=63372) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1092.441516] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-3367136f-9582-4a61-9f78-9904a13146be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.451187] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1092.451187] env[63372]: value = "task-1024831" [ 1092.451187] env[63372]: _type = "Task" [ 1092.451187] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.459849] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024831, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.462670] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 83db51f8-2c4b-4ef1-afcb-ecac63bcab9d] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.524048] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024828, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.744250] env[63372]: INFO nova.compute.manager [-] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Took 1.51 seconds to deallocate network for instance. [ 1092.745229] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.745481] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.745702] env[63372]: DEBUG nova.objects.instance [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lazy-loading 'resources' on Instance uuid 829591bc-9686-4c7d-895d-47a3a1e617c4 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1092.792763] env[63372]: DEBUG oslo_concurrency.lockutils [req-6aa432d4-df65-4247-b5aa-71360a6997a3 req-babd6de3-17f5-439c-bbc4-e5ec4d1d215d service nova] Releasing lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.796174] env[63372]: DEBUG oslo_vmware.api [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024829, 'name': ReconfigVM_Task, 'duration_secs': 0.237554} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.796457] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227463', 'volume_id': '2b5f42b6-add6-497e-9592-e48a741bd2f0', 'name': 'volume-2b5f42b6-add6-497e-9592-e48a741bd2f0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a6bd1c51-e702-48e4-b9ab-8d80db1fec9e', 'attached_at': '', 'detached_at': '', 'volume_id': '2b5f42b6-add6-497e-9592-e48a741bd2f0', 'serial': '2b5f42b6-add6-497e-9592-e48a741bd2f0'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1092.961250] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024831, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.966054] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: c38256cc-215c-4bec-94e5-24a1bbef3e10] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.026841] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024828, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.252871] env[63372]: DEBUG oslo_concurrency.lockutils [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.270971] env[63372]: DEBUG nova.compute.manager [req-5387c2bc-9758-46ad-a30c-5de23d94e3fa req-f6ade7df-8bd5-40c1-b284-9483eb6ad736 service nova] [instance: 4a2e53e0-b68c-42be-afff-1eead017679d] Received event network-vif-deleted-7cd17ebc-3698-499b-8a5a-0507f100485c {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1093.393580] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90fd3705-45e6-4c2d-a1fb-617df4222595 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.402489] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90011cf-2325-494f-b56c-b35088323a9b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.438504] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-671161e3-13dd-4407-a99c-7112f7942a99 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.447394] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aca7465-49ee-4736-8603-4f78dc629f33 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.465841] env[63372]: DEBUG nova.compute.provider_tree [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.470600] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 800cb48b-e75e-45e7-94dc-6ad147e6c6c5] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.472519] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024831, 'name': MoveVirtualDisk_Task} progress is 26%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.528555] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024828, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.846038] env[63372]: DEBUG nova.objects.instance [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lazy-loading 'flavor' on Instance uuid a6bd1c51-e702-48e4-b9ab-8d80db1fec9e {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.965037] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024831, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.974399] env[63372]: DEBUG nova.scheduler.client.report [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1093.977921] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: d7b7ee35-5e83-4c62-bd1e-8ec39951c44c] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.028564] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024828, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.351751] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7686c163-239c-4899-a0b5-c85797a0a535 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.812s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.464806] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024831, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.480864] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.735s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.483498] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: a3145c98-e046-4b55-b80a-57e63807bd7a] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.485708] env[63372]: DEBUG oslo_concurrency.lockutils [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.233s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.485957] env[63372]: DEBUG nova.objects.instance [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lazy-loading 'resources' on Instance uuid 4a2e53e0-b68c-42be-afff-1eead017679d {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.518718] env[63372]: INFO nova.scheduler.client.report [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Deleted allocations for instance 829591bc-9686-4c7d-895d-47a3a1e617c4 [ 1094.542157] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024828, 'name': CreateVM_Task} progress is 99%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.966746] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024831, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.988689] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: cbb471c5-8d86-4d8f-8935-f6a77e5144be] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.034742] env[63372]: DEBUG oslo_concurrency.lockutils [None req-fc50a986-bfd9-4dec-b416-4a1af36f4c20 tempest-ServerDiskConfigTestJSON-757767145 tempest-ServerDiskConfigTestJSON-757767145-project-member] Lock "829591bc-9686-4c7d-895d-47a3a1e617c4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.294s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.040110] env[63372]: INFO nova.compute.manager [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Rescuing [ 1095.040401] env[63372]: DEBUG oslo_concurrency.lockutils [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.040560] env[63372]: DEBUG oslo_concurrency.lockutils [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.040731] env[63372]: DEBUG nova.network.neutron [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1095.042826] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024828, 'name': CreateVM_Task, 'duration_secs': 2.959086} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.043021] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9e5749c-a090-4ce9-b796-852438256502] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1095.044028] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.044028] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.044297] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1095.044723] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76ec7f77-f899-413a-9688-efa4d8c7a018 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.054630] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1095.054630] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b56fae-d4a1-aeaa-41e0-9d4dc91a4ade" [ 1095.054630] env[63372]: _type = "Task" [ 1095.054630] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.068249] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b56fae-d4a1-aeaa-41e0-9d4dc91a4ade, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.121817] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd676fd3-1783-4447-ae7e-89cb11350a96 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.129928] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b039ba6-3a56-4b39-926d-de5d4e9e723b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.161445] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3636f9c4-b244-45b9-b949-8cd16df1b405 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.169873] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ab4be9-48c1-483e-a70f-bc62b2845afe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.184952] env[63372]: DEBUG nova.compute.provider_tree [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.464610] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024831, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.694009} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.465030] env[63372]: INFO nova.virt.vmwareapi.ds_util [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_2ff065e3-d625-4a44-9d78-7cfa53184c82/OSTACK_IMG_2ff065e3-d625-4a44-9d78-7cfa53184c82.vmdk to [datastore1] devstack-image-cache_base/6735c751-59a4-497b-9de2-3765dd7aa055/6735c751-59a4-497b-9de2-3765dd7aa055.vmdk. [ 1095.465309] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Cleaning up location [datastore1] OSTACK_IMG_2ff065e3-d625-4a44-9d78-7cfa53184c82 {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1095.465568] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_2ff065e3-d625-4a44-9d78-7cfa53184c82 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1095.465871] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d341661f-345e-48b6-aa26-a2859c57db38 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.473739] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1095.473739] env[63372]: value = "task-1024832" [ 1095.473739] env[63372]: _type = "Task" [ 1095.473739] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.482828] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024832, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.495592] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: b455b16b-3332-43bf-a91e-a350287ba5f3] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.566106] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b56fae-d4a1-aeaa-41e0-9d4dc91a4ade, 'name': SearchDatastore_Task, 'duration_secs': 0.027092} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.566707] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.566872] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1095.567131] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.567286] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.567668] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1095.568251] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea3328a9-4527-4906-bd95-0e9ea3ca260b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.576865] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1095.577252] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1095.578008] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1c902ac-0402-4259-9c27-4a238dd1e2eb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.583938] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1095.583938] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5235dad0-e987-1936-2718-16ea24685d78" [ 1095.583938] env[63372]: _type = "Task" [ 1095.583938] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.595840] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5235dad0-e987-1936-2718-16ea24685d78, 'name': SearchDatastore_Task, 'duration_secs': 0.009451} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.596682] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f6b78d9-4221-4e57-b9cf-387fbc4262a2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.602010] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1095.602010] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]528848f0-4aee-6cae-35be-b94821dc5515" [ 1095.602010] env[63372]: _type = "Task" [ 1095.602010] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.611273] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]528848f0-4aee-6cae-35be-b94821dc5515, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.687608] env[63372]: DEBUG nova.scheduler.client.report [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.922140] env[63372]: DEBUG nova.network.neutron [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Updating instance_info_cache with network_info: [{"id": "0fbd29da-197e-452a-9644-ab461d995ec3", "address": "fa:16:3e:c8:e0:b1", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fbd29da-19", "ovs_interfaceid": "0fbd29da-197e-452a-9644-ab461d995ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.984073] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024832, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.040457} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.984490] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1095.984490] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6735c751-59a4-497b-9de2-3765dd7aa055/6735c751-59a4-497b-9de2-3765dd7aa055.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.984816] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6735c751-59a4-497b-9de2-3765dd7aa055/6735c751-59a4-497b-9de2-3765dd7aa055.vmdk to [datastore1] 28eb1675-b1b8-46a6-873d-5c858b716575/28eb1675-b1b8-46a6-873d-5c858b716575.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1095.985015] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5011033-231e-4dc2-b6a6-cbeca2fb397b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.992410] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1095.992410] env[63372]: value = "task-1024833" [ 1095.992410] env[63372]: _type = "Task" [ 1095.992410] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.001371] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 6beadbab-2cc1-4b69-95c9-e1a0ea11045f] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.003274] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024833, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.112942] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]528848f0-4aee-6cae-35be-b94821dc5515, 'name': SearchDatastore_Task, 'duration_secs': 0.008984} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.113245] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.113533] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] c9e5749c-a090-4ce9-b796-852438256502/c9e5749c-a090-4ce9-b796-852438256502.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1096.113839] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fccf3f2b-ae6a-4221-b49d-388c3249cff2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.122488] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1096.122488] env[63372]: value = "task-1024834" [ 1096.122488] env[63372]: _type = "Task" [ 1096.122488] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.131073] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024834, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.192791] env[63372]: DEBUG oslo_concurrency.lockutils [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.707s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.215409] env[63372]: INFO nova.scheduler.client.report [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleted allocations for instance 4a2e53e0-b68c-42be-afff-1eead017679d [ 1096.425225] env[63372]: DEBUG oslo_concurrency.lockutils [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.507396] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: b7732621-7e58-40a3-b723-5c66df6f74a1] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.509456] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024833, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.636542] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024834, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.724181] env[63372]: DEBUG oslo_concurrency.lockutils [None req-967959bf-3ce6-4352-bc2b-e26a1ce78705 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "4a2e53e0-b68c-42be-afff-1eead017679d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.741s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.965100] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1096.965100] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b7a3df8-dbd3-467b-8a2b-08dd82b339b1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.977402] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1096.977402] env[63372]: value = "task-1024835" [ 1096.977402] env[63372]: _type = "Task" [ 1096.977402] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.989529] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024835, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.006032] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024833, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.011851] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: c7c444cb-0aaa-4486-840a-e48fa255b421] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.135120] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024834, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.489969] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024835, 'name': PowerOffVM_Task, 'duration_secs': 0.26422} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.490252] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1097.491126] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7515f7dd-5f16-486b-bc51-b988d72cf596 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.504143] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024833, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.521729] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 61f7a052-9a42-4df3-a568-eab197fc5c4f] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.525576] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b22ce4-8926-4e83-a643-03119e1e219f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.566193] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1097.566505] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8bb6ae8-b94b-4d89-89f7-8745a5a57c7f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.577104] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1097.577104] env[63372]: value = "task-1024836" [ 1097.577104] env[63372]: _type = "Task" [ 1097.577104] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.590986] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] VM already powered off {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1097.591259] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1097.591546] env[63372]: DEBUG oslo_concurrency.lockutils [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.591704] env[63372]: DEBUG oslo_concurrency.lockutils [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.591935] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1097.592241] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-90797d9d-d47f-4e32-b341-ea25d522cce1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.612072] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1097.612348] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1097.613162] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16ab2f56-e616-4f12-80b5-dcbc0e70fcab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.624163] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1097.624163] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5284d0b6-0fd5-127d-204b-04633d05d9de" [ 1097.624163] env[63372]: _type = "Task" [ 1097.624163] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.639637] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5284d0b6-0fd5-127d-204b-04633d05d9de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.644620] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024834, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.007440] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024833, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.025236] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 4ee6c0ac-b7ac-49e7-b029-ca6e9a595912] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.137348] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024834, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.146365] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5284d0b6-0fd5-127d-204b-04633d05d9de, 'name': SearchDatastore_Task, 'duration_secs': 0.092354} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.146365] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12d39f60-278a-4c13-a00b-b9944ad7f073 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.152711] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1098.152711] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]526bed7a-01e8-a061-d1f8-0e3e2d41f9ca" [ 1098.152711] env[63372]: _type = "Task" [ 1098.152711] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.162687] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526bed7a-01e8-a061-d1f8-0e3e2d41f9ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.372647] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "802deb8b-978b-4d01-b19c-a0090e0029f0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.373033] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "802deb8b-978b-4d01-b19c-a0090e0029f0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.506424] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024833, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.529124] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: ed2cefa1-3604-43c5-9860-0fb49133bbb4] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.634264] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024834, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.36491} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.634504] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] c9e5749c-a090-4ce9-b796-852438256502/c9e5749c-a090-4ce9-b796-852438256502.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1098.634723] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1098.634977] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b1c5daaf-94fc-44f6-a3fe-abc71604b1d6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.642859] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1098.642859] env[63372]: value = "task-1024837" [ 1098.642859] env[63372]: _type = "Task" [ 1098.642859] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.652832] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024837, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.662872] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]526bed7a-01e8-a061-d1f8-0e3e2d41f9ca, 'name': SearchDatastore_Task, 'duration_secs': 0.122912} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.663181] env[63372]: DEBUG oslo_concurrency.lockutils [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.663449] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] a6bd1c51-e702-48e4-b9ab-8d80db1fec9e/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk. {{(pid=63372) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1098.664124] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1d70397-a3cd-43a4-9bd1-90edbead2178 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.670782] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1098.670782] env[63372]: value = "task-1024838" [ 1098.670782] env[63372]: _type = "Task" [ 1098.670782] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.678853] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024838, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.876338] env[63372]: DEBUG nova.compute.manager [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1099.007522] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024833, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.560033} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.007893] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6735c751-59a4-497b-9de2-3765dd7aa055/6735c751-59a4-497b-9de2-3765dd7aa055.vmdk to [datastore1] 28eb1675-b1b8-46a6-873d-5c858b716575/28eb1675-b1b8-46a6-873d-5c858b716575.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1099.008956] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45de033d-dff1-4737-9522-b2c15bf3df15 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.036364] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 28eb1675-b1b8-46a6-873d-5c858b716575/28eb1675-b1b8-46a6-873d-5c858b716575.vmdk or device None with type streamOptimized {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1099.036884] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 91985614-b959-401e-bb06-d67b230ee026] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.038757] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aaba4a70-e6b1-479b-a984-0c98e4f9b0b9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.061408] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1099.061408] env[63372]: value = "task-1024839" [ 1099.061408] env[63372]: _type = "Task" [ 1099.061408] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.078214] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024839, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.154343] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024837, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073905} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.154625] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1099.155405] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b3530b-8539-4024-b909-884f3f5f4905 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.179242] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] c9e5749c-a090-4ce9-b796-852438256502/c9e5749c-a090-4ce9-b796-852438256502.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1099.179434] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2890038a-a56d-44da-968b-e96080e86425 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.207310] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1099.207310] env[63372]: value = "task-1024840" [ 1099.207310] env[63372]: _type = "Task" [ 1099.207310] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.211031] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024838, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.4548} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.215079] env[63372]: INFO nova.virt.vmwareapi.ds_util [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] a6bd1c51-e702-48e4-b9ab-8d80db1fec9e/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk. [ 1099.216262] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d4c1c5-e943-47cb-9a9c-d40cb13d1ffa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.230888] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024840, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.255710] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] a6bd1c51-e702-48e4-b9ab-8d80db1fec9e/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1099.256060] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68ad5696-66bb-4712-9ae3-5d0a9f766839 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.277732] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1099.277732] env[63372]: value = "task-1024841" [ 1099.277732] env[63372]: _type = "Task" [ 1099.277732] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.290130] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024841, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.401684] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.402059] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.404726] env[63372]: INFO nova.compute.claims [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1099.554173] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 95178e2e-9d71-4606-a856-c530bfbb9345] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.573242] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.721018] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024840, 'name': ReconfigVM_Task, 'duration_secs': 0.323644} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.721663] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Reconfigured VM instance instance-0000006d to attach disk [datastore1] c9e5749c-a090-4ce9-b796-852438256502/c9e5749c-a090-4ce9-b796-852438256502.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1099.722325] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-63a2d359-8602-45ea-b49c-72878f3f1714 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.730375] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1099.730375] env[63372]: value = "task-1024842" [ 1099.730375] env[63372]: _type = "Task" [ 1099.730375] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.739798] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024842, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.789750] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024841, 'name': ReconfigVM_Task, 'duration_secs': 0.355007} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.790046] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Reconfigured VM instance instance-00000068 to attach disk [datastore1] a6bd1c51-e702-48e4-b9ab-8d80db1fec9e/d7ae1717-77b2-47f6-9acd-b27bac221f7c-rescue.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1099.792032] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea1ce94-0656-4ddb-9821-6469515a930c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.824728] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b397213-a2cb-40a3-a1dd-14393239f5ba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.843528] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1099.843528] env[63372]: value = "task-1024843" [ 1099.843528] env[63372]: _type = "Task" [ 1099.843528] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.861036] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024843, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.059400] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 662d9fdf-ec0d-4466-8d15-abc942e4dad8] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.077572] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024839, 'name': ReconfigVM_Task, 'duration_secs': 0.577763} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.077572] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 28eb1675-b1b8-46a6-873d-5c858b716575/28eb1675-b1b8-46a6-873d-5c858b716575.vmdk or device None with type streamOptimized {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1100.077572] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af9d8380-6de7-4814-8441-2d0f61a746c3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.083958] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1100.083958] env[63372]: value = "task-1024844" [ 1100.083958] env[63372]: _type = "Task" [ 1100.083958] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.097843] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024844, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.241053] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024842, 'name': Rename_Task, 'duration_secs': 0.166507} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.241410] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1100.241666] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-faf5ff8f-c971-43e0-9020-2109bc042e52 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.253020] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1100.253020] env[63372]: value = "task-1024845" [ 1100.253020] env[63372]: _type = "Task" [ 1100.253020] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.262585] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024845, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.359054] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024843, 'name': ReconfigVM_Task, 'duration_secs': 0.183861} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.359054] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1100.359054] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94284f06-4ab1-46af-9338-f5be8e865e50 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.365603] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1100.365603] env[63372]: value = "task-1024846" [ 1100.365603] env[63372]: _type = "Task" [ 1100.365603] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.377184] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024846, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.561474] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 326a9155-1c00-4c18-a9a7-6124e709c20b] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.596042] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024844, 'name': Rename_Task, 'duration_secs': 0.143649} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.598759] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1100.599884] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d87a62f-a6d4-4943-9d4d-ef42519bce9e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.606483] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1100.606483] env[63372]: value = "task-1024847" [ 1100.606483] env[63372]: _type = "Task" [ 1100.606483] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.617712] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024847, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.640553] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f3fdf9-746d-48b0-97b7-172ced2ce443 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.652360] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a64d11-1338-4ff5-a935-0631bdcc2f8b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.684800] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-863c7d32-6af8-44b7-b7b6-08664b89a363 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.693312] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d02e0d-a03d-4309-a8f4-212de3440c27 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.709887] env[63372]: DEBUG nova.compute.provider_tree [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.772384] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024845, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.880228] env[63372]: DEBUG oslo_vmware.api [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024846, 'name': PowerOnVM_Task, 'duration_secs': 0.468171} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.880548] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1100.883422] env[63372]: DEBUG nova.compute.manager [None req-36d6d8f6-f6fa-47d2-98d7-f0a6d8ce1cee tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1100.884249] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d152fe9c-08dd-4658-acb7-9e1cb17d7a28 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.067699] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: d568b727-7836-4b1d-9f27-f159227e46f7] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.117166] env[63372]: DEBUG oslo_vmware.api [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024847, 'name': PowerOnVM_Task, 'duration_secs': 0.490523} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.117475] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1101.214777] env[63372]: DEBUG nova.scheduler.client.report [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1101.218398] env[63372]: DEBUG nova.compute.manager [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1101.218776] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d39c0562-0255-4a2d-9f33-57297c71079a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.264329] env[63372]: DEBUG oslo_vmware.api [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024845, 'name': PowerOnVM_Task, 'duration_secs': 0.54835} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.264592] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1101.264793] env[63372]: INFO nova.compute.manager [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Took 11.94 seconds to spawn the instance on the hypervisor. [ 1101.264973] env[63372]: DEBUG nova.compute.manager [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1101.265965] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433cd953-ca7e-416b-9e41-347fe876bf3b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.571789] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: ba1daa0a-dcf8-4586-9964-a962fcc94196] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.617401] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "fafed45e-c9dc-4196-883f-a23570fd0583" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.618355] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.723369] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.321s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.723910] env[63372]: DEBUG nova.compute.manager [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1101.738797] env[63372]: DEBUG oslo_concurrency.lockutils [None req-78528e57-3c8f-4ae9-bfcf-d9081e8e4c16 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "28eb1675-b1b8-46a6-873d-5c858b716575" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 26.299s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.782867] env[63372]: INFO nova.compute.manager [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Took 22.71 seconds to build instance. [ 1102.076539] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 75560852-8c6d-4220-b251-81e8721a4b7b] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.119668] env[63372]: DEBUG nova.compute.manager [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1102.156545] env[63372]: INFO nova.compute.manager [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Unrescuing [ 1102.156822] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.157072] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquired lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.157165] env[63372]: DEBUG nova.network.neutron [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1102.229087] env[63372]: DEBUG nova.compute.utils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1102.230274] env[63372]: DEBUG nova.compute.manager [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1102.230907] env[63372]: DEBUG nova.network.neutron [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1102.270069] env[63372]: DEBUG nova.policy [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d822682f06640489a48f8577acf4389', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e46b6acae4a4b71bc390bd513ddd2f1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1102.284607] env[63372]: DEBUG oslo_concurrency.lockutils [None req-98b2a929-20cb-4e11-b54e-37a15b022d8f tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c9e5749c-a090-4ce9-b796-852438256502" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.224s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.579681] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 382fa76b-b930-459e-b84e-da1b9eb74b74] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.618830] env[63372]: DEBUG nova.network.neutron [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Successfully created port: 146031c4-2f32-4085-9fc6-3060e7db9d08 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1102.647085] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.647710] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.649042] env[63372]: INFO nova.compute.claims [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1102.737016] env[63372]: DEBUG nova.compute.manager [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1102.947786] env[63372]: DEBUG nova.network.neutron [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Updating instance_info_cache with network_info: [{"id": "0fbd29da-197e-452a-9644-ab461d995ec3", "address": "fa:16:3e:c8:e0:b1", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fbd29da-19", "ovs_interfaceid": "0fbd29da-197e-452a-9644-ab461d995ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.082649] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: c0b9bd7e-92e2-44a1-b461-e54c4411ddc8] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.450796] env[63372]: DEBUG oslo_concurrency.lockutils [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Releasing lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.452332] env[63372]: DEBUG nova.objects.instance [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lazy-loading 'flavor' on Instance uuid a6bd1c51-e702-48e4-b9ab-8d80db1fec9e {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.534186] env[63372]: DEBUG nova.compute.manager [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Stashing vm_state: active {{(pid=63372) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1103.586344] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 1f52f9ee-2af3-4321-8ec6-5d5e841ed880] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.746607] env[63372]: DEBUG nova.compute.manager [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1103.776217] env[63372]: DEBUG nova.virt.hardware [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1103.776500] env[63372]: DEBUG nova.virt.hardware [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1103.776667] env[63372]: DEBUG nova.virt.hardware [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1103.776883] env[63372]: DEBUG nova.virt.hardware [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1103.777048] env[63372]: DEBUG nova.virt.hardware [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1103.777222] env[63372]: DEBUG nova.virt.hardware [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1103.777452] env[63372]: DEBUG nova.virt.hardware [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1103.777614] env[63372]: DEBUG nova.virt.hardware [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1103.777808] env[63372]: DEBUG nova.virt.hardware [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1103.777980] env[63372]: DEBUG nova.virt.hardware [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1103.778193] env[63372]: DEBUG nova.virt.hardware [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1103.779151] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb626eb5-f339-4140-b5c6-13cbe290f996 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.791128] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70507022-28b1-4420-8df8-69cc952bcc14 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.810498] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c101ce2-9ea7-4ee8-b463-91cf922bea73 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.818206] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c545211-1da0-464e-bbe5-0a929be71897 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.849209] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90339d86-1997-4aef-b06e-18fbe8e15b16 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.857885] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5fbe27-f4c2-4e6d-8624-1c2e4b16fc70 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.872510] env[63372]: DEBUG nova.compute.provider_tree [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.958779] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caedbdc7-ecbf-45eb-8476-af2113ae72e3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.986562] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1103.986941] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d708cc8-ca89-407f-afcd-1d8dca405f60 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.999008] env[63372]: DEBUG oslo_vmware.api [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1103.999008] env[63372]: value = "task-1024848" [ 1103.999008] env[63372]: _type = "Task" [ 1103.999008] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.009973] env[63372]: DEBUG oslo_vmware.api [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024848, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.045397] env[63372]: DEBUG nova.compute.manager [req-01497d12-c7b1-49bd-9335-1ec88d97c3aa req-3b0d3fe8-7092-4c5a-9721-974288aba668 service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Received event network-vif-plugged-146031c4-2f32-4085-9fc6-3060e7db9d08 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1104.046332] env[63372]: DEBUG oslo_concurrency.lockutils [req-01497d12-c7b1-49bd-9335-1ec88d97c3aa req-3b0d3fe8-7092-4c5a-9721-974288aba668 service nova] Acquiring lock "802deb8b-978b-4d01-b19c-a0090e0029f0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.046576] env[63372]: DEBUG oslo_concurrency.lockutils [req-01497d12-c7b1-49bd-9335-1ec88d97c3aa req-3b0d3fe8-7092-4c5a-9721-974288aba668 service nova] Lock "802deb8b-978b-4d01-b19c-a0090e0029f0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.046766] env[63372]: DEBUG oslo_concurrency.lockutils [req-01497d12-c7b1-49bd-9335-1ec88d97c3aa req-3b0d3fe8-7092-4c5a-9721-974288aba668 service nova] Lock "802deb8b-978b-4d01-b19c-a0090e0029f0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.047042] env[63372]: DEBUG nova.compute.manager [req-01497d12-c7b1-49bd-9335-1ec88d97c3aa req-3b0d3fe8-7092-4c5a-9721-974288aba668 service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] No waiting events found dispatching network-vif-plugged-146031c4-2f32-4085-9fc6-3060e7db9d08 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1104.047152] env[63372]: WARNING nova.compute.manager [req-01497d12-c7b1-49bd-9335-1ec88d97c3aa req-3b0d3fe8-7092-4c5a-9721-974288aba668 service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Received unexpected event network-vif-plugged-146031c4-2f32-4085-9fc6-3060e7db9d08 for instance with vm_state building and task_state spawning. [ 1104.053537] env[63372]: DEBUG oslo_concurrency.lockutils [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.089601] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: da8cf1ff-d977-4fd1-9239-043f07bed875] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.158777] env[63372]: DEBUG nova.network.neutron [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Successfully updated port: 146031c4-2f32-4085-9fc6-3060e7db9d08 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1104.375589] env[63372]: DEBUG nova.scheduler.client.report [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1104.509987] env[63372]: DEBUG oslo_vmware.api [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024848, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.592861] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 13bde897-8446-42a2-b02d-2f5b48e6f432] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.663076] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.663076] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.663076] env[63372]: DEBUG nova.network.neutron [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1104.881098] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.233s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.881226] env[63372]: DEBUG nova.compute.manager [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1104.883943] env[63372]: DEBUG oslo_concurrency.lockutils [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.830s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.012260] env[63372]: DEBUG oslo_vmware.api [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024848, 'name': PowerOffVM_Task, 'duration_secs': 0.553173} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.012574] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1105.018318] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Reconfiguring VM instance instance-00000068 to detach disk 2002 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1105.018662] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c681aee2-49a3-4678-af85-1b303bd62235 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.038322] env[63372]: DEBUG oslo_vmware.api [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1105.038322] env[63372]: value = "task-1024849" [ 1105.038322] env[63372]: _type = "Task" [ 1105.038322] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.047665] env[63372]: DEBUG oslo_vmware.api [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024849, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.095921] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 1d047728-50d8-465c-b217-6cbe9bbea3e8] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.198590] env[63372]: DEBUG nova.network.neutron [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1105.341295] env[63372]: DEBUG nova.network.neutron [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Updating instance_info_cache with network_info: [{"id": "146031c4-2f32-4085-9fc6-3060e7db9d08", "address": "fa:16:3e:70:3b:d5", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap146031c4-2f", "ovs_interfaceid": "146031c4-2f32-4085-9fc6-3060e7db9d08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.386636] env[63372]: DEBUG nova.compute.utils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1105.388428] env[63372]: DEBUG nova.compute.manager [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1105.388600] env[63372]: DEBUG nova.network.neutron [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1105.393406] env[63372]: INFO nova.compute.claims [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1105.437125] env[63372]: DEBUG nova.policy [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '228dd399ae7c45ad964fa74c0cbb044c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4b088060a2c4159a62e40e7e627076a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1105.552913] env[63372]: DEBUG oslo_vmware.api [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024849, 'name': ReconfigVM_Task, 'duration_secs': 0.272373} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.554048] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Reconfigured VM instance instance-00000068 to detach disk 2002 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1105.554440] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1105.554870] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66c4ee9a-e37f-44a5-ba96-ccf55fdeb63c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.564906] env[63372]: DEBUG oslo_vmware.api [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1105.564906] env[63372]: value = "task-1024850" [ 1105.564906] env[63372]: _type = "Task" [ 1105.564906] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.578271] env[63372]: DEBUG oslo_vmware.api [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024850, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.600097] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 798047e4-1a2b-458c-86c2-5567197d1861] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.734023] env[63372]: DEBUG nova.network.neutron [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Successfully created port: 7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1105.843863] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.844503] env[63372]: DEBUG nova.compute.manager [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Instance network_info: |[{"id": "146031c4-2f32-4085-9fc6-3060e7db9d08", "address": "fa:16:3e:70:3b:d5", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap146031c4-2f", "ovs_interfaceid": "146031c4-2f32-4085-9fc6-3060e7db9d08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1105.844711] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:3b:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d5d0e0d-cdec-474a-a891-a9ceff15a8b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '146031c4-2f32-4085-9fc6-3060e7db9d08', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1105.852673] env[63372]: DEBUG oslo.service.loopingcall [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1105.852823] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1105.853022] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4e28acdd-d395-4ab5-ae5f-b131c692b7a6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.876503] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1105.876503] env[63372]: value = "task-1024851" [ 1105.876503] env[63372]: _type = "Task" [ 1105.876503] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.885613] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024851, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.900194] env[63372]: INFO nova.compute.resource_tracker [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating resource usage from migration 3bf0c6bd-830c-4d5d-9efa-aa9bd8e71e97 [ 1105.903388] env[63372]: DEBUG nova.compute.manager [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1106.075089] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cdbfb4d-d26f-4d9d-a5d5-263e5f2fe5a1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.080433] env[63372]: DEBUG nova.compute.manager [req-b4cfad91-ed1f-4d05-8d5f-7b49ebe3d4d5 req-e96edb2d-3a84-4c8f-83cf-b348b271ac55 service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Received event network-changed-146031c4-2f32-4085-9fc6-3060e7db9d08 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1106.080642] env[63372]: DEBUG nova.compute.manager [req-b4cfad91-ed1f-4d05-8d5f-7b49ebe3d4d5 req-e96edb2d-3a84-4c8f-83cf-b348b271ac55 service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Refreshing instance network info cache due to event network-changed-146031c4-2f32-4085-9fc6-3060e7db9d08. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1106.080858] env[63372]: DEBUG oslo_concurrency.lockutils [req-b4cfad91-ed1f-4d05-8d5f-7b49ebe3d4d5 req-e96edb2d-3a84-4c8f-83cf-b348b271ac55 service nova] Acquiring lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.081019] env[63372]: DEBUG oslo_concurrency.lockutils [req-b4cfad91-ed1f-4d05-8d5f-7b49ebe3d4d5 req-e96edb2d-3a84-4c8f-83cf-b348b271ac55 service nova] Acquired lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.081211] env[63372]: DEBUG nova.network.neutron [req-b4cfad91-ed1f-4d05-8d5f-7b49ebe3d4d5 req-e96edb2d-3a84-4c8f-83cf-b348b271ac55 service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Refreshing network info cache for port 146031c4-2f32-4085-9fc6-3060e7db9d08 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1106.086386] env[63372]: DEBUG oslo_vmware.api [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024850, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.092184] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a4e2f9-1342-4ded-b7fe-e58d7ede8f9c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.125362] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 07dab964-a741-4723-98a3-0cfbe4eff975] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.128133] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3b67a54-14fd-48dd-8d89-a94f45b12f01 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.137060] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bac398e-b617-433a-9aab-e535bd3e53d7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.154271] env[63372]: DEBUG nova.compute.provider_tree [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.387133] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024851, 'name': CreateVM_Task, 'duration_secs': 0.395681} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.387286] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1106.388265] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.388521] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.388896] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1106.389199] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d1fa3df-7493-4768-9539-a7936d3e553a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.394716] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1106.394716] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5235fe2e-5fda-e432-7b6f-902ee2091b8e" [ 1106.394716] env[63372]: _type = "Task" [ 1106.394716] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.403669] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5235fe2e-5fda-e432-7b6f-902ee2091b8e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.576113] env[63372]: DEBUG oslo_vmware.api [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024850, 'name': PowerOnVM_Task, 'duration_secs': 0.539652} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.576401] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1106.576633] env[63372]: DEBUG nova.compute.manager [None req-5d63757a-80e9-4d8c-92b3-89247d5f9bd0 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1106.577401] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79474251-9ada-4ed6-a778-0f779b500cf1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.627875] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: a3d5b93e-1b76-48e7-bb3e-2423e7c4119a] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.656695] env[63372]: DEBUG nova.scheduler.client.report [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1106.788010] env[63372]: DEBUG nova.network.neutron [req-b4cfad91-ed1f-4d05-8d5f-7b49ebe3d4d5 req-e96edb2d-3a84-4c8f-83cf-b348b271ac55 service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Updated VIF entry in instance network info cache for port 146031c4-2f32-4085-9fc6-3060e7db9d08. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1106.788402] env[63372]: DEBUG nova.network.neutron [req-b4cfad91-ed1f-4d05-8d5f-7b49ebe3d4d5 req-e96edb2d-3a84-4c8f-83cf-b348b271ac55 service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Updating instance_info_cache with network_info: [{"id": "146031c4-2f32-4085-9fc6-3060e7db9d08", "address": "fa:16:3e:70:3b:d5", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap146031c4-2f", "ovs_interfaceid": "146031c4-2f32-4085-9fc6-3060e7db9d08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.907063] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5235fe2e-5fda-e432-7b6f-902ee2091b8e, 'name': SearchDatastore_Task, 'duration_secs': 0.01212} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.907570] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.907906] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1106.908246] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.908464] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.908724] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1106.909081] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ad507ac1-7e96-4971-9798-8e43a8f1ff06 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.914336] env[63372]: DEBUG nova.compute.manager [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1106.921126] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1106.921389] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1106.922467] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-013c1ae7-214a-4050-b3b4-14c6f0bf89e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.931230] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1106.931230] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523a10a7-39f1-1e3c-946c-abbe44a45760" [ 1106.931230] env[63372]: _type = "Task" [ 1106.931230] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.945262] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523a10a7-39f1-1e3c-946c-abbe44a45760, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.948921] env[63372]: DEBUG nova.virt.hardware [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1106.949292] env[63372]: DEBUG nova.virt.hardware [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1106.949519] env[63372]: DEBUG nova.virt.hardware [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1106.949783] env[63372]: DEBUG nova.virt.hardware [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1106.949992] env[63372]: DEBUG nova.virt.hardware [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1106.950216] env[63372]: DEBUG nova.virt.hardware [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1106.950502] env[63372]: DEBUG nova.virt.hardware [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1106.951265] env[63372]: DEBUG nova.virt.hardware [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1106.951265] env[63372]: DEBUG nova.virt.hardware [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1106.951265] env[63372]: DEBUG nova.virt.hardware [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1106.951550] env[63372]: DEBUG nova.virt.hardware [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1106.953032] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30e5799-fe08-490c-aafd-133f6aa81f7f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.960760] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde290c3-1d60-4b01-af7b-a05a9453cb3d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.130785] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 4028d7c0-f398-4e43-9cff-5d89a14c4efa] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1107.163878] env[63372]: DEBUG oslo_concurrency.lockutils [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.280s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.164132] env[63372]: INFO nova.compute.manager [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Migrating [ 1107.230652] env[63372]: DEBUG nova.network.neutron [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Successfully updated port: 7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1107.291584] env[63372]: DEBUG oslo_concurrency.lockutils [req-b4cfad91-ed1f-4d05-8d5f-7b49ebe3d4d5 req-e96edb2d-3a84-4c8f-83cf-b348b271ac55 service nova] Releasing lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.443149] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523a10a7-39f1-1e3c-946c-abbe44a45760, 'name': SearchDatastore_Task, 'duration_secs': 0.026923} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.443974] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49d8a0de-9553-4410-a087-67857dd571ca {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.451040] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1107.451040] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523705e1-a851-9e4f-c28e-e4d190a2c7f8" [ 1107.451040] env[63372]: _type = "Task" [ 1107.451040] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.458973] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523705e1-a851-9e4f-c28e-e4d190a2c7f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.633997] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 89301344-84a9-4d13-aae7-99943d0a478e] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1107.678895] env[63372]: DEBUG oslo_concurrency.lockutils [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.679057] env[63372]: DEBUG oslo_concurrency.lockutils [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.679316] env[63372]: DEBUG nova.network.neutron [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1107.733244] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.733508] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquired lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.733790] env[63372]: DEBUG nova.network.neutron [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1107.941511] env[63372]: DEBUG nova.compute.manager [req-439748e1-4c4c-48d0-8cab-15ab5347119f req-e3e0cd78-5dd0-417a-9995-5cd08a5b383d service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Received event network-changed-0fbd29da-197e-452a-9644-ab461d995ec3 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1107.941644] env[63372]: DEBUG nova.compute.manager [req-439748e1-4c4c-48d0-8cab-15ab5347119f req-e3e0cd78-5dd0-417a-9995-5cd08a5b383d service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Refreshing instance network info cache due to event network-changed-0fbd29da-197e-452a-9644-ab461d995ec3. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1107.942344] env[63372]: DEBUG oslo_concurrency.lockutils [req-439748e1-4c4c-48d0-8cab-15ab5347119f req-e3e0cd78-5dd0-417a-9995-5cd08a5b383d service nova] Acquiring lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.942524] env[63372]: DEBUG oslo_concurrency.lockutils [req-439748e1-4c4c-48d0-8cab-15ab5347119f req-e3e0cd78-5dd0-417a-9995-5cd08a5b383d service nova] Acquired lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.942697] env[63372]: DEBUG nova.network.neutron [req-439748e1-4c4c-48d0-8cab-15ab5347119f req-e3e0cd78-5dd0-417a-9995-5cd08a5b383d service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Refreshing network info cache for port 0fbd29da-197e-452a-9644-ab461d995ec3 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1107.962711] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523705e1-a851-9e4f-c28e-e4d190a2c7f8, 'name': SearchDatastore_Task, 'duration_secs': 0.009869} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.962981] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.964037] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 802deb8b-978b-4d01-b19c-a0090e0029f0/802deb8b-978b-4d01-b19c-a0090e0029f0.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1107.964037] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-005d3ada-6802-4d2f-b800-f80005496b2b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.970661] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1107.970661] env[63372]: value = "task-1024852" [ 1107.970661] env[63372]: _type = "Task" [ 1107.970661] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.980326] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024852, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.111575] env[63372]: DEBUG nova.compute.manager [req-ae429be9-80a7-434b-8e90-7ac578f05301 req-2e28a911-281c-469b-a752-d96d30a8c511 service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Received event network-vif-plugged-7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1108.111846] env[63372]: DEBUG oslo_concurrency.lockutils [req-ae429be9-80a7-434b-8e90-7ac578f05301 req-2e28a911-281c-469b-a752-d96d30a8c511 service nova] Acquiring lock "fafed45e-c9dc-4196-883f-a23570fd0583-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.112124] env[63372]: DEBUG oslo_concurrency.lockutils [req-ae429be9-80a7-434b-8e90-7ac578f05301 req-2e28a911-281c-469b-a752-d96d30a8c511 service nova] Lock "fafed45e-c9dc-4196-883f-a23570fd0583-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.112353] env[63372]: DEBUG oslo_concurrency.lockutils [req-ae429be9-80a7-434b-8e90-7ac578f05301 req-2e28a911-281c-469b-a752-d96d30a8c511 service nova] Lock "fafed45e-c9dc-4196-883f-a23570fd0583-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.112646] env[63372]: DEBUG nova.compute.manager [req-ae429be9-80a7-434b-8e90-7ac578f05301 req-2e28a911-281c-469b-a752-d96d30a8c511 service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] No waiting events found dispatching network-vif-plugged-7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1108.112872] env[63372]: WARNING nova.compute.manager [req-ae429be9-80a7-434b-8e90-7ac578f05301 req-2e28a911-281c-469b-a752-d96d30a8c511 service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Received unexpected event network-vif-plugged-7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db for instance with vm_state building and task_state spawning. [ 1108.113113] env[63372]: DEBUG nova.compute.manager [req-ae429be9-80a7-434b-8e90-7ac578f05301 req-2e28a911-281c-469b-a752-d96d30a8c511 service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Received event network-changed-7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1108.113343] env[63372]: DEBUG nova.compute.manager [req-ae429be9-80a7-434b-8e90-7ac578f05301 req-2e28a911-281c-469b-a752-d96d30a8c511 service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Refreshing instance network info cache due to event network-changed-7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1108.113587] env[63372]: DEBUG oslo_concurrency.lockutils [req-ae429be9-80a7-434b-8e90-7ac578f05301 req-2e28a911-281c-469b-a752-d96d30a8c511 service nova] Acquiring lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.136962] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 368101ee-e4fd-4fe4-b7c8-75c6b210189f] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.278730] env[63372]: DEBUG nova.network.neutron [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1108.430290] env[63372]: DEBUG nova.network.neutron [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating instance_info_cache with network_info: [{"id": "4a95ce9d-92de-44b7-96b7-59f4238835c7", "address": "fa:16:3e:51:87:f7", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a95ce9d-92", "ovs_interfaceid": "4a95ce9d-92de-44b7-96b7-59f4238835c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.442373] env[63372]: DEBUG nova.network.neutron [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Updating instance_info_cache with network_info: [{"id": "7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db", "address": "fa:16:3e:94:43:d7", "network": {"id": "b161f533-292b-431a-a171-e9d61bc90105", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1928954423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4b088060a2c4159a62e40e7e627076a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ca80379-c6", "ovs_interfaceid": "7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.483141] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024852, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457338} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.483461] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 802deb8b-978b-4d01-b19c-a0090e0029f0/802deb8b-978b-4d01-b19c-a0090e0029f0.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1108.483680] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1108.483947] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a451685c-5a6f-4db6-8b7c-d12f407ac196 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.491721] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1108.491721] env[63372]: value = "task-1024853" [ 1108.491721] env[63372]: _type = "Task" [ 1108.491721] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.500084] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024853, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.639853] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 357505d0-f306-4e11-8a62-e03cfab2b7c5] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.744137] env[63372]: DEBUG nova.network.neutron [req-439748e1-4c4c-48d0-8cab-15ab5347119f req-e3e0cd78-5dd0-417a-9995-5cd08a5b383d service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Updated VIF entry in instance network info cache for port 0fbd29da-197e-452a-9644-ab461d995ec3. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1108.744589] env[63372]: DEBUG nova.network.neutron [req-439748e1-4c4c-48d0-8cab-15ab5347119f req-e3e0cd78-5dd0-417a-9995-5cd08a5b383d service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Updating instance_info_cache with network_info: [{"id": "0fbd29da-197e-452a-9644-ab461d995ec3", "address": "fa:16:3e:c8:e0:b1", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fbd29da-19", "ovs_interfaceid": "0fbd29da-197e-452a-9644-ab461d995ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.936297] env[63372]: DEBUG oslo_concurrency.lockutils [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.944922] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Releasing lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.945347] env[63372]: DEBUG nova.compute.manager [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Instance network_info: |[{"id": "7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db", "address": "fa:16:3e:94:43:d7", "network": {"id": "b161f533-292b-431a-a171-e9d61bc90105", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1928954423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4b088060a2c4159a62e40e7e627076a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ca80379-c6", "ovs_interfaceid": "7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1108.945680] env[63372]: DEBUG oslo_concurrency.lockutils [req-ae429be9-80a7-434b-8e90-7ac578f05301 req-2e28a911-281c-469b-a752-d96d30a8c511 service nova] Acquired lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.945874] env[63372]: DEBUG nova.network.neutron [req-ae429be9-80a7-434b-8e90-7ac578f05301 req-2e28a911-281c-469b-a752-d96d30a8c511 service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Refreshing network info cache for port 7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1108.947229] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:43:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8480e3f5-68bd-4c27-ae1f-7c994a8202b1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1108.955474] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Creating folder: Project (f4b088060a2c4159a62e40e7e627076a). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1108.956733] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-558b3d18-52f1-42c8-b3d6-f5ea716a5bb5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.969026] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Created folder: Project (f4b088060a2c4159a62e40e7e627076a) in parent group-v227230. [ 1108.969026] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Creating folder: Instances. Parent ref: group-v227468. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1108.969026] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7be5917f-0b1f-403e-b3cc-bf8e309015d4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.979897] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Created folder: Instances in parent group-v227468. [ 1108.980499] env[63372]: DEBUG oslo.service.loopingcall [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1108.980499] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1108.980682] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fd20a3af-a432-461b-bed1-2fe70f865dc4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.004872] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024853, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.006643] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1109.006643] env[63372]: value = "task-1024856" [ 1109.006643] env[63372]: _type = "Task" [ 1109.006643] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.014012] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024856, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.142868] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: c01a5d24-eb46-4a69-993e-753880ce8e85] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1109.247417] env[63372]: DEBUG oslo_concurrency.lockutils [req-439748e1-4c4c-48d0-8cab-15ab5347119f req-e3e0cd78-5dd0-417a-9995-5cd08a5b383d service nova] Releasing lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.504808] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024853, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.000828} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.505158] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1109.505843] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c571322-3de8-4769-a717-fb00fc8723c9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.517962] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024856, 'name': CreateVM_Task, 'duration_secs': 0.376246} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.526447] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1109.535119] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 802deb8b-978b-4d01-b19c-a0090e0029f0/802deb8b-978b-4d01-b19c-a0090e0029f0.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1109.537712] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.537877] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.538201] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1109.538431] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-925d4cf7-2a1d-453a-9a07-fd8435c486ec {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.552044] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c73d597-2998-4c0b-8cb7-ec0db2fe9730 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.558540] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1109.558540] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52600b9b-bb75-ddcf-4e50-2cf6dd1e4139" [ 1109.558540] env[63372]: _type = "Task" [ 1109.558540] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.562752] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1109.562752] env[63372]: value = "task-1024857" [ 1109.562752] env[63372]: _type = "Task" [ 1109.562752] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.568657] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52600b9b-bb75-ddcf-4e50-2cf6dd1e4139, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.575385] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024857, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.645977] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 7c9d551e-d210-4943-971f-b2829751fcfc] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1109.681907] env[63372]: DEBUG nova.network.neutron [req-ae429be9-80a7-434b-8e90-7ac578f05301 req-2e28a911-281c-469b-a752-d96d30a8c511 service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Updated VIF entry in instance network info cache for port 7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1109.682291] env[63372]: DEBUG nova.network.neutron [req-ae429be9-80a7-434b-8e90-7ac578f05301 req-2e28a911-281c-469b-a752-d96d30a8c511 service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Updating instance_info_cache with network_info: [{"id": "7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db", "address": "fa:16:3e:94:43:d7", "network": {"id": "b161f533-292b-431a-a171-e9d61bc90105", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1928954423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4b088060a2c4159a62e40e7e627076a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ca80379-c6", "ovs_interfaceid": "7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.968216] env[63372]: DEBUG nova.compute.manager [req-88974608-7c51-47eb-a664-31cdaa5af6c2 req-ca168f51-af7e-4acb-adc6-39121b8c0b1e service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Received event network-changed-0fbd29da-197e-452a-9644-ab461d995ec3 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1109.968377] env[63372]: DEBUG nova.compute.manager [req-88974608-7c51-47eb-a664-31cdaa5af6c2 req-ca168f51-af7e-4acb-adc6-39121b8c0b1e service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Refreshing instance network info cache due to event network-changed-0fbd29da-197e-452a-9644-ab461d995ec3. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1109.968593] env[63372]: DEBUG oslo_concurrency.lockutils [req-88974608-7c51-47eb-a664-31cdaa5af6c2 req-ca168f51-af7e-4acb-adc6-39121b8c0b1e service nova] Acquiring lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.968801] env[63372]: DEBUG oslo_concurrency.lockutils [req-88974608-7c51-47eb-a664-31cdaa5af6c2 req-ca168f51-af7e-4acb-adc6-39121b8c0b1e service nova] Acquired lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.969040] env[63372]: DEBUG nova.network.neutron [req-88974608-7c51-47eb-a664-31cdaa5af6c2 req-ca168f51-af7e-4acb-adc6-39121b8c0b1e service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Refreshing network info cache for port 0fbd29da-197e-452a-9644-ab461d995ec3 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1110.070021] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52600b9b-bb75-ddcf-4e50-2cf6dd1e4139, 'name': SearchDatastore_Task, 'duration_secs': 0.010386} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.070627] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.070912] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1110.071123] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.071273] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.071483] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1110.071746] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-078a0293-54fb-4721-a1ec-0100c508963e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.076328] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024857, 'name': ReconfigVM_Task, 'duration_secs': 0.280149} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.076882] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 802deb8b-978b-4d01-b19c-a0090e0029f0/802deb8b-978b-4d01-b19c-a0090e0029f0.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1110.077481] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d62c2694-5a10-4896-b61d-9631c97eeb54 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.083741] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1110.083741] env[63372]: value = "task-1024858" [ 1110.083741] env[63372]: _type = "Task" [ 1110.083741] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.084477] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1110.084653] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1110.087930] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4ae78d6-d3cb-46f9-b40e-9c57fa4f0d1f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.095975] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1110.095975] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529f2d86-aa2c-a52d-b459-4ffe65679faa" [ 1110.095975] env[63372]: _type = "Task" [ 1110.095975] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.096208] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024858, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.103593] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529f2d86-aa2c-a52d-b459-4ffe65679faa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.148822] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: de1b38a4-c7f2-420c-a050-7311976e4ca8] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.184486] env[63372]: DEBUG oslo_concurrency.lockutils [req-ae429be9-80a7-434b-8e90-7ac578f05301 req-2e28a911-281c-469b-a752-d96d30a8c511 service nova] Releasing lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.452182] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd3dc85-b0b6-48ea-b831-54fbaccd88ee {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.474867] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating instance 'c9e5749c-a090-4ce9-b796-852438256502' progress to 0 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1110.593690] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024858, 'name': Rename_Task, 'duration_secs': 0.176398} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.593988] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1110.594253] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ef49b6e-9f5d-485b-9719-d7de3bb992fc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.604129] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1110.604129] env[63372]: value = "task-1024859" [ 1110.604129] env[63372]: _type = "Task" [ 1110.604129] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.607757] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529f2d86-aa2c-a52d-b459-4ffe65679faa, 'name': SearchDatastore_Task, 'duration_secs': 0.014236} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.611196] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11e776a3-f4e8-45c0-95f5-b25918b679cd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.621108] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1110.621108] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52ce5df6-cdcb-9604-8037-1006c2505882" [ 1110.621108] env[63372]: _type = "Task" [ 1110.621108] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.624922] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024859, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.635863] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ce5df6-cdcb-9604-8037-1006c2505882, 'name': SearchDatastore_Task, 'duration_secs': 0.009654} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.636189] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.636559] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] fafed45e-c9dc-4196-883f-a23570fd0583/fafed45e-c9dc-4196-883f-a23570fd0583.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1110.637253] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f2c59db-844b-4e22-a9ec-c6d24494e343 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.644631] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1110.644631] env[63372]: value = "task-1024860" [ 1110.644631] env[63372]: _type = "Task" [ 1110.644631] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.653198] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 99f901a6-9bb3-4403-af0c-c8900f655cb3] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.655060] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024860, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.706425] env[63372]: DEBUG nova.network.neutron [req-88974608-7c51-47eb-a664-31cdaa5af6c2 req-ca168f51-af7e-4acb-adc6-39121b8c0b1e service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Updated VIF entry in instance network info cache for port 0fbd29da-197e-452a-9644-ab461d995ec3. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1110.707026] env[63372]: DEBUG nova.network.neutron [req-88974608-7c51-47eb-a664-31cdaa5af6c2 req-ca168f51-af7e-4acb-adc6-39121b8c0b1e service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Updating instance_info_cache with network_info: [{"id": "0fbd29da-197e-452a-9644-ab461d995ec3", "address": "fa:16:3e:c8:e0:b1", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fbd29da-19", "ovs_interfaceid": "0fbd29da-197e-452a-9644-ab461d995ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.981478] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1110.981950] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b3fcf47-e0e9-4b99-abdb-e0fcb5207e74 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.992631] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1110.992631] env[63372]: value = "task-1024861" [ 1110.992631] env[63372]: _type = "Task" [ 1110.992631] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.002672] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024861, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.119714] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024859, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.156704] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: f6675697-1529-46be-b28a-398ff3060d18] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.158634] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024860, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468038} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.159340] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] fafed45e-c9dc-4196-883f-a23570fd0583/fafed45e-c9dc-4196-883f-a23570fd0583.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1111.159728] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1111.160131] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b2929248-a8e4-498f-a24a-62cbc827be09 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.170976] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1111.170976] env[63372]: value = "task-1024862" [ 1111.170976] env[63372]: _type = "Task" [ 1111.170976] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.180151] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024862, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.210409] env[63372]: DEBUG oslo_concurrency.lockutils [req-88974608-7c51-47eb-a664-31cdaa5af6c2 req-ca168f51-af7e-4acb-adc6-39121b8c0b1e service nova] Releasing lock "refresh_cache-a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.502990] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024861, 'name': PowerOffVM_Task, 'duration_secs': 0.27374} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.503297] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1111.503488] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating instance 'c9e5749c-a090-4ce9-b796-852438256502' progress to 17 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1111.618769] env[63372]: DEBUG oslo_vmware.api [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024859, 'name': PowerOnVM_Task, 'duration_secs': 0.588947} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.619055] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1111.619257] env[63372]: INFO nova.compute.manager [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Took 7.87 seconds to spawn the instance on the hypervisor. [ 1111.619445] env[63372]: DEBUG nova.compute.manager [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1111.620223] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2ee06c-1ed6-4a48-97d5-802efa1202cf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.660205] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 3f66eda6-f5e9-4527-9711-849a01702580] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.680495] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024862, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083513} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.680763] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1111.681581] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e9a0991-0bfa-4d5e-a6cd-d17547ea68a5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.704462] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] fafed45e-c9dc-4196-883f-a23570fd0583/fafed45e-c9dc-4196-883f-a23570fd0583.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1111.704919] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4aeb8c66-8f8d-4e11-8f8f-ff79063ef2d2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.724401] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1111.724401] env[63372]: value = "task-1024863" [ 1111.724401] env[63372]: _type = "Task" [ 1111.724401] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.732103] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024863, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.010227] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1112.010498] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1112.010662] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1112.010844] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1112.010990] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1112.011156] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1112.011358] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1112.011602] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1112.011808] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1112.011978] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1112.012173] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1112.017139] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39fd325f-6cd3-4994-8bae-289979d8d026 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.035215] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1112.035215] env[63372]: value = "task-1024864" [ 1112.035215] env[63372]: _type = "Task" [ 1112.035215] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.045105] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024864, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.139192] env[63372]: INFO nova.compute.manager [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Took 12.76 seconds to build instance. [ 1112.163277] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 69a107d0-80c7-42e5-b514-b4273e1a3359] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1112.235797] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024863, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.547313] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024864, 'name': ReconfigVM_Task, 'duration_secs': 0.388795} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.549276] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating instance 'c9e5749c-a090-4ce9-b796-852438256502' progress to 33 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1112.641480] env[63372]: DEBUG oslo_concurrency.lockutils [None req-0c166c96-d917-4cbc-93ed-514351d48383 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "802deb8b-978b-4d01-b19c-a0090e0029f0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.268s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.668097] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: ae6f3f42-7213-4ab1-b74c-1a557df6748b] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1112.734917] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024863, 'name': ReconfigVM_Task, 'duration_secs': 0.978709} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.737739] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Reconfigured VM instance instance-0000006f to attach disk [datastore2] fafed45e-c9dc-4196-883f-a23570fd0583/fafed45e-c9dc-4196-883f-a23570fd0583.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1112.737739] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4cd689df-b225-4417-beab-2f0983928a16 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.745019] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1112.745019] env[63372]: value = "task-1024865" [ 1112.745019] env[63372]: _type = "Task" [ 1112.745019] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.753242] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024865, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.809523] env[63372]: DEBUG nova.compute.manager [req-d5a1c754-0c92-46d0-a16e-5fa472f8bef3 req-63d0af85-3f04-44e0-bffb-516b7efeb30f service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Received event network-changed-146031c4-2f32-4085-9fc6-3060e7db9d08 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.809523] env[63372]: DEBUG nova.compute.manager [req-d5a1c754-0c92-46d0-a16e-5fa472f8bef3 req-63d0af85-3f04-44e0-bffb-516b7efeb30f service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Refreshing instance network info cache due to event network-changed-146031c4-2f32-4085-9fc6-3060e7db9d08. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1112.809721] env[63372]: DEBUG oslo_concurrency.lockutils [req-d5a1c754-0c92-46d0-a16e-5fa472f8bef3 req-63d0af85-3f04-44e0-bffb-516b7efeb30f service nova] Acquiring lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.809879] env[63372]: DEBUG oslo_concurrency.lockutils [req-d5a1c754-0c92-46d0-a16e-5fa472f8bef3 req-63d0af85-3f04-44e0-bffb-516b7efeb30f service nova] Acquired lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.810058] env[63372]: DEBUG nova.network.neutron [req-d5a1c754-0c92-46d0-a16e-5fa472f8bef3 req-63d0af85-3f04-44e0-bffb-516b7efeb30f service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Refreshing network info cache for port 146031c4-2f32-4085-9fc6-3060e7db9d08 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1113.058941] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1113.059261] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1113.059496] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1113.059745] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1113.059939] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1113.060146] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1113.060413] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1113.060700] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1113.060939] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1113.061170] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1113.061454] env[63372]: DEBUG nova.virt.hardware [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1113.069207] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Reconfiguring VM instance instance-0000006d to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1113.069753] env[63372]: DEBUG oslo_concurrency.lockutils [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.070049] env[63372]: DEBUG oslo_concurrency.lockutils [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.070266] env[63372]: INFO nova.compute.manager [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Shelving [ 1113.071847] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-073de38b-103e-4635-9688-be6cd3f83b90 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.100154] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1113.100468] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea07e031-2a28-4faf-9654-70ba4d5de33e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.106190] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1113.106190] env[63372]: value = "task-1024866" [ 1113.106190] env[63372]: _type = "Task" [ 1113.106190] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.111558] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1113.111558] env[63372]: value = "task-1024867" [ 1113.111558] env[63372]: _type = "Task" [ 1113.111558] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.118172] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024866, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.122950] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.169292] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: cd0c01ac-602b-44a3-8099-84b8a50b2449] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.254295] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024865, 'name': Rename_Task, 'duration_secs': 0.193589} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.254572] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1113.254789] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ed422d4-a976-4d54-8716-bd4913cd1c4f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.262133] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1113.262133] env[63372]: value = "task-1024868" [ 1113.262133] env[63372]: _type = "Task" [ 1113.262133] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.276917] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024868, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.627011] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024866, 'name': ReconfigVM_Task, 'duration_secs': 0.206567} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.627345] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Reconfigured VM instance instance-0000006d to detach disk 2000 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1113.628525] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e34aeb-f813-4460-ac8c-93004da65e23 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.641484] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.666927] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] c9e5749c-a090-4ce9-b796-852438256502/c9e5749c-a090-4ce9-b796-852438256502.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1113.669683] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cae4511b-0e92-4267-80e5-f24e72c56a14 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.683314] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: f43555ef-b517-4b7d-9d2e-4787d40e201b] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.693552] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1113.693552] env[63372]: value = "task-1024869" [ 1113.693552] env[63372]: _type = "Task" [ 1113.693552] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.702804] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024869, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.780078] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024868, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.843205] env[63372]: DEBUG nova.network.neutron [req-d5a1c754-0c92-46d0-a16e-5fa472f8bef3 req-63d0af85-3f04-44e0-bffb-516b7efeb30f service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Updated VIF entry in instance network info cache for port 146031c4-2f32-4085-9fc6-3060e7db9d08. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1113.843626] env[63372]: DEBUG nova.network.neutron [req-d5a1c754-0c92-46d0-a16e-5fa472f8bef3 req-63d0af85-3f04-44e0-bffb-516b7efeb30f service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Updating instance_info_cache with network_info: [{"id": "146031c4-2f32-4085-9fc6-3060e7db9d08", "address": "fa:16:3e:70:3b:d5", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap146031c4-2f", "ovs_interfaceid": "146031c4-2f32-4085-9fc6-3060e7db9d08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.123170] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.186859] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 93a5d948-0629-4f53-a681-858d519acfa7] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1114.204104] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024869, 'name': ReconfigVM_Task, 'duration_secs': 0.334846} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.204423] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Reconfigured VM instance instance-0000006d to attach disk [datastore1] c9e5749c-a090-4ce9-b796-852438256502/c9e5749c-a090-4ce9-b796-852438256502.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1114.204709] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating instance 'c9e5749c-a090-4ce9-b796-852438256502' progress to 50 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1114.276202] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024868, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.346325] env[63372]: DEBUG oslo_concurrency.lockutils [req-d5a1c754-0c92-46d0-a16e-5fa472f8bef3 req-63d0af85-3f04-44e0-bffb-516b7efeb30f service nova] Releasing lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.623038] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.690928] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 8283b736-ad02-4082-97b7-561bd5c5da93] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1114.711100] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed87fd98-0283-45ef-9ffc-3d342245b6aa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.732210] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30887c87-ebeb-4157-bc33-14f02f939f2b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.751218] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating instance 'c9e5749c-a090-4ce9-b796-852438256502' progress to 67 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1114.776777] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024868, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.127269] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.195137] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: f0c60559-c072-4b61-afe8-03d6c131b307] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1115.277538] env[63372]: DEBUG oslo_vmware.api [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024868, 'name': PowerOnVM_Task, 'duration_secs': 1.521128} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.277769] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1115.277956] env[63372]: INFO nova.compute.manager [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Took 8.36 seconds to spawn the instance on the hypervisor. [ 1115.278154] env[63372]: DEBUG nova.compute.manager [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1115.278953] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef4aa3f-a09c-430f-9576-c22f4c4edefc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.291396] env[63372]: DEBUG nova.network.neutron [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Port 4a95ce9d-92de-44b7-96b7-59f4238835c7 binding to destination host cpu-1 is already ACTIVE {{(pid=63372) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1115.625519] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.698351] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: b182294d-2de8-4189-af7f-3e2d2c604a8b] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1115.802886] env[63372]: INFO nova.compute.manager [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Took 13.18 seconds to build instance. [ 1115.900417] env[63372]: DEBUG nova.compute.manager [req-499961ed-3a75-405e-a27f-e89b1b767dfa req-b64a19c4-5f47-4e39-97f0-651582d5e77a service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Received event network-changed-7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1115.900417] env[63372]: DEBUG nova.compute.manager [req-499961ed-3a75-405e-a27f-e89b1b767dfa req-b64a19c4-5f47-4e39-97f0-651582d5e77a service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Refreshing instance network info cache due to event network-changed-7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1115.900417] env[63372]: DEBUG oslo_concurrency.lockutils [req-499961ed-3a75-405e-a27f-e89b1b767dfa req-b64a19c4-5f47-4e39-97f0-651582d5e77a service nova] Acquiring lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.900417] env[63372]: DEBUG oslo_concurrency.lockutils [req-499961ed-3a75-405e-a27f-e89b1b767dfa req-b64a19c4-5f47-4e39-97f0-651582d5e77a service nova] Acquired lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.900576] env[63372]: DEBUG nova.network.neutron [req-499961ed-3a75-405e-a27f-e89b1b767dfa req-b64a19c4-5f47-4e39-97f0-651582d5e77a service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Refreshing network info cache for port 7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1116.124756] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024867, 'name': PowerOffVM_Task, 'duration_secs': 2.722441} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.125135] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1116.125804] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936a322a-5235-450f-82d3-e26642bc4986 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.143948] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1bfe94-5989-423c-adf7-a9c51a14d18b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.201551] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 5d819bd2-f7f9-480d-83ae-fbd4d2e0d382] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1116.308659] env[63372]: DEBUG oslo_concurrency.lockutils [None req-89624937-0cda-4898-99c7-6e8d65681a0c tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.691s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.314943] env[63372]: DEBUG oslo_concurrency.lockutils [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "c9e5749c-a090-4ce9-b796-852438256502-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.314943] env[63372]: DEBUG oslo_concurrency.lockutils [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c9e5749c-a090-4ce9-b796-852438256502-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.315134] env[63372]: DEBUG oslo_concurrency.lockutils [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c9e5749c-a090-4ce9-b796-852438256502-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.620919] env[63372]: DEBUG nova.network.neutron [req-499961ed-3a75-405e-a27f-e89b1b767dfa req-b64a19c4-5f47-4e39-97f0-651582d5e77a service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Updated VIF entry in instance network info cache for port 7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1116.621368] env[63372]: DEBUG nova.network.neutron [req-499961ed-3a75-405e-a27f-e89b1b767dfa req-b64a19c4-5f47-4e39-97f0-651582d5e77a service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Updating instance_info_cache with network_info: [{"id": "7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db", "address": "fa:16:3e:94:43:d7", "network": {"id": "b161f533-292b-431a-a171-e9d61bc90105", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1928954423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4b088060a2c4159a62e40e7e627076a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ca80379-c6", "ovs_interfaceid": "7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.656200] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Creating Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1116.656466] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-692968fe-3e77-4b4f-8325-60872eb055a7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.664902] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1116.664902] env[63372]: value = "task-1024870" [ 1116.664902] env[63372]: _type = "Task" [ 1116.664902] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.673610] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024870, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.704721] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: d50919ac-3a0b-46ac-a837-ca3e6ad05173] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1117.123851] env[63372]: DEBUG oslo_concurrency.lockutils [req-499961ed-3a75-405e-a27f-e89b1b767dfa req-b64a19c4-5f47-4e39-97f0-651582d5e77a service nova] Releasing lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.175818] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024870, 'name': CreateSnapshot_Task, 'duration_secs': 0.421596} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.176178] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Created Snapshot of the VM instance {{(pid=63372) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1117.176995] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99bd42bc-572c-450c-ba13-6e58af759dfa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.208344] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 19be9aeb-d534-4ab2-8f9b-3147d2c9f7a5] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1117.353496] env[63372]: DEBUG oslo_concurrency.lockutils [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1117.353697] env[63372]: DEBUG oslo_concurrency.lockutils [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.353876] env[63372]: DEBUG nova.network.neutron [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1117.695748] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Creating linked-clone VM from snapshot {{(pid=63372) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1117.696086] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f978b2a3-4abf-4b9a-8082-14a1388d59c5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.705566] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1117.705566] env[63372]: value = "task-1024871" [ 1117.705566] env[63372]: _type = "Task" [ 1117.705566] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.713923] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: c4718797-aa86-4ec0-94d3-6480bd6aa898] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1117.715696] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024871, 'name': CloneVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.055580] env[63372]: DEBUG nova.network.neutron [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating instance_info_cache with network_info: [{"id": "4a95ce9d-92de-44b7-96b7-59f4238835c7", "address": "fa:16:3e:51:87:f7", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a95ce9d-92", "ovs_interfaceid": "4a95ce9d-92de-44b7-96b7-59f4238835c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.215609] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024871, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.217035] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 5819c38e-2cf9-4d16-b28a-5f23d35c3d44] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1118.559100] env[63372]: DEBUG oslo_concurrency.lockutils [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.716097] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024871, 'name': CloneVM_Task} progress is 94%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.720606] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: ac90a156-be00-4f62-a76e-e08914531167] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1119.098127] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac71d544-c361-4133-89ef-7cec243fcbba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.119138] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ff9165-7163-4b1a-b2fe-33c44885e0a8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.126474] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating instance 'c9e5749c-a090-4ce9-b796-852438256502' progress to 83 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1119.215922] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024871, 'name': CloneVM_Task} progress is 95%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.223476] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: bbba9f28-045d-41ab-8539-5b2968fe3d54] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1119.632904] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1119.633203] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6385d0f-14d5-4353-80ff-1766f0974fed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.641574] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1119.641574] env[63372]: value = "task-1024872" [ 1119.641574] env[63372]: _type = "Task" [ 1119.641574] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.649291] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024872, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.717118] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024871, 'name': CloneVM_Task, 'duration_secs': 1.740684} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.717416] env[63372]: INFO nova.virt.vmwareapi.vmops [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Created linked-clone VM from snapshot [ 1119.718205] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7699e0c4-1372-4a05-b266-ee5db353f18f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.725691] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Uploading image 53b2f81e-8df7-4596-909e-8ed3e82311c7 {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1119.729491] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: cf673ac1-2c7d-468b-83ec-c723d5182457] Instance has had 0 of 5 cleanup attempts {{(pid=63372) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1119.761656] env[63372]: DEBUG oslo_vmware.rw_handles [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1119.761656] env[63372]: value = "vm-227472" [ 1119.761656] env[63372]: _type = "VirtualMachine" [ 1119.761656] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1119.761930] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8678e00b-5199-440b-812c-0b9f9a91f5a8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.769610] env[63372]: DEBUG oslo_vmware.rw_handles [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lease: (returnval){ [ 1119.769610] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5287cc38-71ee-b759-fe19-f441f9d4aede" [ 1119.769610] env[63372]: _type = "HttpNfcLease" [ 1119.769610] env[63372]: } obtained for exporting VM: (result){ [ 1119.769610] env[63372]: value = "vm-227472" [ 1119.769610] env[63372]: _type = "VirtualMachine" [ 1119.769610] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1119.769865] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the lease: (returnval){ [ 1119.769865] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5287cc38-71ee-b759-fe19-f441f9d4aede" [ 1119.769865] env[63372]: _type = "HttpNfcLease" [ 1119.769865] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1119.776759] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1119.776759] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5287cc38-71ee-b759-fe19-f441f9d4aede" [ 1119.776759] env[63372]: _type = "HttpNfcLease" [ 1119.776759] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1120.154242] env[63372]: DEBUG oslo_vmware.api [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024872, 'name': PowerOnVM_Task, 'duration_secs': 0.379514} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.154547] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1120.154738] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-939add49-18f4-45df-b53a-854493b72d91 tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating instance 'c9e5749c-a090-4ce9-b796-852438256502' progress to 100 {{(pid=63372) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1120.232931] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.232931] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Cleaning up deleted instances with incomplete migration {{(pid=63372) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1120.277950] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1120.277950] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5287cc38-71ee-b759-fe19-f441f9d4aede" [ 1120.277950] env[63372]: _type = "HttpNfcLease" [ 1120.277950] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1120.278266] env[63372]: DEBUG oslo_vmware.rw_handles [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1120.278266] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5287cc38-71ee-b759-fe19-f441f9d4aede" [ 1120.278266] env[63372]: _type = "HttpNfcLease" [ 1120.278266] env[63372]: }. {{(pid=63372) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1120.278974] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc1e862-c0df-4697-871d-2a4ffa5655e2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.286054] env[63372]: DEBUG oslo_vmware.rw_handles [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527077c9-ae44-eabf-18f3-90ece6d0b4f7/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1120.286238] env[63372]: DEBUG oslo_vmware.rw_handles [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527077c9-ae44-eabf-18f3-90ece6d0b4f7/disk-0.vmdk for reading. {{(pid=63372) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1120.380938] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e2ad1f93-1fc6-47d9-a1e5-9dc51612a517 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.735038] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.018876] env[63372]: DEBUG oslo_concurrency.lockutils [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.019234] env[63372]: DEBUG oslo_concurrency.lockutils [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.186696] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "c9e5749c-a090-4ce9-b796-852438256502" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.187168] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c9e5749c-a090-4ce9-b796-852438256502" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.187427] env[63372]: DEBUG nova.compute.manager [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Going to confirm migration 8 {{(pid=63372) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1122.233918] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.523394] env[63372]: DEBUG nova.compute.utils [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1122.740826] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._sync_power_states {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.760990] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.761214] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquired lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.761395] env[63372]: DEBUG nova.network.neutron [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1122.761643] env[63372]: DEBUG nova.objects.instance [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lazy-loading 'info_cache' on Instance uuid c9e5749c-a090-4ce9-b796-852438256502 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1123.027413] env[63372]: DEBUG oslo_concurrency.lockutils [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.244955] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Getting list of instances from cluster (obj){ [ 1123.244955] env[63372]: value = "domain-c8" [ 1123.244955] env[63372]: _type = "ClusterComputeResource" [ 1123.244955] env[63372]: } {{(pid=63372) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1123.246109] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25b1af6-de5f-4eea-abff-18ed6fdde02e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.266782] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Got total of 9 instances {{(pid=63372) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1123.267368] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Triggering sync for uuid ddca7aa5-fcc8-4835-bfdb-47781335f2cd {{(pid=63372) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1123.267611] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Triggering sync for uuid 1724ce03-c8e2-415d-a380-59ac69fbfb57 {{(pid=63372) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1123.267800] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Triggering sync for uuid 28eb1675-b1b8-46a6-873d-5c858b716575 {{(pid=63372) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1123.268096] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Triggering sync for uuid a6bd1c51-e702-48e4-b9ab-8d80db1fec9e {{(pid=63372) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1123.268276] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Triggering sync for uuid 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b {{(pid=63372) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1123.268428] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Triggering sync for uuid 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e {{(pid=63372) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1123.268697] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Triggering sync for uuid c9e5749c-a090-4ce9-b796-852438256502 {{(pid=63372) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1123.268832] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Triggering sync for uuid 802deb8b-978b-4d01-b19c-a0090e0029f0 {{(pid=63372) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1123.268912] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Triggering sync for uuid fafed45e-c9dc-4196-883f-a23570fd0583 {{(pid=63372) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1123.269521] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.269764] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.270046] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "1724ce03-c8e2-415d-a380-59ac69fbfb57" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.270261] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "1724ce03-c8e2-415d-a380-59ac69fbfb57" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.270502] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "28eb1675-b1b8-46a6-873d-5c858b716575" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.270725] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "28eb1675-b1b8-46a6-873d-5c858b716575" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.270976] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.271183] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.271409] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.271669] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.271857] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.272098] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "c9e5749c-a090-4ce9-b796-852438256502" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.272306] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "802deb8b-978b-4d01-b19c-a0090e0029f0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.272518] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "802deb8b-978b-4d01-b19c-a0090e0029f0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.273075] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "fafed45e-c9dc-4196-883f-a23570fd0583" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.273291] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.274708] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b60d526c-c404-44c6-985d-de50fed9018e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.279289] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95865646-61eb-48c0-8fe4-58071c47f879 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.282108] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39414e1b-9c5c-4cf9-baec-2bef74c53fe4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.284956] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb6ac79-b6bc-431a-b1ba-7f5bf5b751c9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.287659] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b2c0624-77c9-48b2-85a4-d3d491425641 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.290393] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a59a5287-0d14-4399-af50-3b9e6ff8344a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.293892] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fe4484-a9c6-4d30-a49b-10828d2fdb24 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.803904] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.533s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.810476] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "28eb1675-b1b8-46a6-873d-5c858b716575" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.540s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.811973] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "1724ce03-c8e2-415d-a380-59ac69fbfb57" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.542s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.813409] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.544s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.814826] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.543s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.831546] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "802deb8b-978b-4d01-b19c-a0090e0029f0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.559s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.831977] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.559s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.956203] env[63372]: DEBUG nova.network.neutron [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating instance_info_cache with network_info: [{"id": "4a95ce9d-92de-44b7-96b7-59f4238835c7", "address": "fa:16:3e:51:87:f7", "network": {"id": "a0107f83-f809-47f5-a462-c8e43f2677b4", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1087342382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe39cfedf214d50be775ef736f94da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a95ce9d-92", "ovs_interfaceid": "4a95ce9d-92de-44b7-96b7-59f4238835c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.092406] env[63372]: DEBUG oslo_concurrency.lockutils [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.092780] env[63372]: DEBUG oslo_concurrency.lockutils [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.092887] env[63372]: INFO nova.compute.manager [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Attaching volume a7250982-c18c-4394-a135-fdca2a2eb278 to /dev/sdb [ 1124.123934] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2a909d-9e32-47f7-b144-9c272d036c56 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.130866] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09412e73-01d3-4525-a550-00cd7dffd006 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.144531] env[63372]: DEBUG nova.virt.block_device [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Updating existing volume attachment record: 4919e4a0-36a4-409e-8ca1-df844b06ee39 {{(pid=63372) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1124.458668] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Releasing lock "refresh_cache-c9e5749c-a090-4ce9-b796-852438256502" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.458980] env[63372]: DEBUG nova.objects.instance [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lazy-loading 'migration_context' on Instance uuid c9e5749c-a090-4ce9-b796-852438256502 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1124.961943] env[63372]: DEBUG nova.objects.base [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1124.963179] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422f1091-dd1b-498c-914a-2a53cef73d47 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.986834] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93dd1e49-9f5d-4553-8f27-98296129349a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.993109] env[63372]: DEBUG oslo_vmware.api [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1124.993109] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]524033c0-f234-0695-a75f-f5db74814012" [ 1124.993109] env[63372]: _type = "Task" [ 1124.993109] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.003959] env[63372]: DEBUG oslo_vmware.api [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524033c0-f234-0695-a75f-f5db74814012, 'name': SearchDatastore_Task, 'duration_secs': 0.007887} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.004268] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.004505] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.631561] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98c4dcc-e9c1-4e8b-acc1-d72d35205c64 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.639075] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebaa2326-cb31-4ee0-be6f-4435a7d16e87 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.668900] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d28eb0-79ed-4604-bc21-57e0361d9ff1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.676314] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04d9f1e-2b28-4bf6-9697-af64958c4e95 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.689336] env[63372]: DEBUG nova.compute.provider_tree [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1126.192443] env[63372]: DEBUG nova.scheduler.client.report [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1127.203063] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.198s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.756568] env[63372]: INFO nova.scheduler.client.report [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleted allocation for migration 3bf0c6bd-830c-4d5d-9efa-aa9bd8e71e97 [ 1128.262944] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c9e5749c-a090-4ce9-b796-852438256502" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.075s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.263576] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "c9e5749c-a090-4ce9-b796-852438256502" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.991s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.263779] env[63372]: INFO nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: c9e5749c-a090-4ce9-b796-852438256502] During sync_power_state the instance has a pending task (deleting). Skip. [ 1128.263983] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "c9e5749c-a090-4ce9-b796-852438256502" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.687019] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Volume attach. Driver type: vmdk {{(pid=63372) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1128.687279] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227474', 'volume_id': 'a7250982-c18c-4394-a135-fdca2a2eb278', 'name': 'volume-a7250982-c18c-4394-a135-fdca2a2eb278', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e', 'attached_at': '', 'detached_at': '', 'volume_id': 'a7250982-c18c-4394-a135-fdca2a2eb278', 'serial': 'a7250982-c18c-4394-a135-fdca2a2eb278'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1128.688211] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13aeadc-2fa4-48e8-9958-51650856ed0d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.704939] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ed4e88-e4e7-49d8-8d70-900c29044df0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.728982] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] volume-a7250982-c18c-4394-a135-fdca2a2eb278/volume-a7250982-c18c-4394-a135-fdca2a2eb278.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1128.729303] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d16c7ca3-90f7-47b7-a888-ff9243efc197 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.748471] env[63372]: DEBUG oslo_vmware.api [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1128.748471] env[63372]: value = "task-1024878" [ 1128.748471] env[63372]: _type = "Task" [ 1128.748471] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.759027] env[63372]: DEBUG oslo_vmware.api [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024878, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.259092] env[63372]: DEBUG oslo_vmware.api [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024878, 'name': ReconfigVM_Task, 'duration_secs': 0.438151} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.259439] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Reconfigured VM instance instance-0000006b to attach disk [datastore1] volume-a7250982-c18c-4394-a135-fdca2a2eb278/volume-a7250982-c18c-4394-a135-fdca2a2eb278.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1129.264382] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-595f6062-813a-466a-9f67-1b4b0680b760 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.279645] env[63372]: DEBUG oslo_vmware.api [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1129.279645] env[63372]: value = "task-1024879" [ 1129.279645] env[63372]: _type = "Task" [ 1129.279645] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.288388] env[63372]: DEBUG oslo_vmware.api [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024879, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.706496] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "c9e5749c-a090-4ce9-b796-852438256502" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.706757] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c9e5749c-a090-4ce9-b796-852438256502" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.706967] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "c9e5749c-a090-4ce9-b796-852438256502-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.707174] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c9e5749c-a090-4ce9-b796-852438256502-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.707348] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c9e5749c-a090-4ce9-b796-852438256502-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.709443] env[63372]: INFO nova.compute.manager [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Terminating instance [ 1129.712460] env[63372]: DEBUG nova.compute.manager [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1129.712460] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1129.714302] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698c3917-3e49-4b68-af65-91af4be40b35 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.721074] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1129.721307] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f863dd37-cf78-4e52-aaef-4e3a40394b4f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.727853] env[63372]: DEBUG oslo_vmware.api [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1129.727853] env[63372]: value = "task-1024880" [ 1129.727853] env[63372]: _type = "Task" [ 1129.727853] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.736751] env[63372]: DEBUG oslo_vmware.api [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024880, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.790237] env[63372]: DEBUG oslo_vmware.api [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024879, 'name': ReconfigVM_Task, 'duration_secs': 0.164308} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.790543] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227474', 'volume_id': 'a7250982-c18c-4394-a135-fdca2a2eb278', 'name': 'volume-a7250982-c18c-4394-a135-fdca2a2eb278', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e', 'attached_at': '', 'detached_at': '', 'volume_id': 'a7250982-c18c-4394-a135-fdca2a2eb278', 'serial': 'a7250982-c18c-4394-a135-fdca2a2eb278'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1129.793597] env[63372]: DEBUG oslo_vmware.rw_handles [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527077c9-ae44-eabf-18f3-90ece6d0b4f7/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1129.794528] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb1e359-dead-4552-bda2-1268a13d34f0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.800920] env[63372]: DEBUG oslo_vmware.rw_handles [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527077c9-ae44-eabf-18f3-90ece6d0b4f7/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1129.801102] env[63372]: ERROR oslo_vmware.rw_handles [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527077c9-ae44-eabf-18f3-90ece6d0b4f7/disk-0.vmdk due to incomplete transfer. [ 1129.802076] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-61bd46e6-bf3c-4ff2-95b9-fd4e934838df {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.811454] env[63372]: DEBUG oslo_vmware.rw_handles [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527077c9-ae44-eabf-18f3-90ece6d0b4f7/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1129.811692] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Uploaded image 53b2f81e-8df7-4596-909e-8ed3e82311c7 to the Glance image server {{(pid=63372) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1129.814928] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Destroying the VM {{(pid=63372) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1129.815461] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f50590ad-6578-41f1-8c16-eacdfe654e0d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.821386] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1129.821386] env[63372]: value = "task-1024881" [ 1129.821386] env[63372]: _type = "Task" [ 1129.821386] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.830258] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024881, 'name': Destroy_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.238160] env[63372]: DEBUG oslo_vmware.api [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024880, 'name': PowerOffVM_Task, 'duration_secs': 0.180825} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.238439] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1130.238608] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1130.238855] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6bb50e22-b1de-4244-8f66-ce1f6bbfa599 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.331731] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024881, 'name': Destroy_Task, 'duration_secs': 0.377629} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.333688] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Destroyed the VM [ 1130.333941] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Deleting Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1130.334386] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-705f45e3-91ec-4866-bdcd-d2e9b0611a21 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.341011] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1130.341011] env[63372]: value = "task-1024883" [ 1130.341011] env[63372]: _type = "Task" [ 1130.341011] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.349477] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024883, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.367573] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1130.367797] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1130.367981] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleting the datastore file [datastore1] c9e5749c-a090-4ce9-b796-852438256502 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1130.368259] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f9b22a20-7dc4-4399-9521-4d467c221c70 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.374152] env[63372]: DEBUG oslo_vmware.api [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for the task: (returnval){ [ 1130.374152] env[63372]: value = "task-1024884" [ 1130.374152] env[63372]: _type = "Task" [ 1130.374152] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.382544] env[63372]: DEBUG oslo_vmware.api [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024884, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.835596] env[63372]: DEBUG nova.objects.instance [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lazy-loading 'flavor' on Instance uuid 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.850648] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024883, 'name': RemoveSnapshot_Task, 'duration_secs': 0.374934} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.851022] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Deleted Snapshot of the VM instance {{(pid=63372) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1130.851300] env[63372]: DEBUG nova.compute.manager [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1130.852142] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb34338-b096-41f9-ae34-6efb1538b10a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.883349] env[63372]: DEBUG oslo_vmware.api [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Task: {'id': task-1024884, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147624} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.883637] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1130.883833] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1130.884037] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1130.884199] env[63372]: INFO nova.compute.manager [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] [instance: c9e5749c-a090-4ce9-b796-852438256502] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1130.884442] env[63372]: DEBUG oslo.service.loopingcall [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1130.884636] env[63372]: DEBUG nova.compute.manager [-] [instance: c9e5749c-a090-4ce9-b796-852438256502] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1130.884732] env[63372]: DEBUG nova.network.neutron [-] [instance: c9e5749c-a090-4ce9-b796-852438256502] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1131.118525] env[63372]: DEBUG nova.compute.manager [req-29ce3d70-56c3-488a-97c6-6f2853d21619 req-01e8a652-031e-4c6c-9d7f-f993280f868a service nova] [instance: c9e5749c-a090-4ce9-b796-852438256502] Received event network-vif-deleted-4a95ce9d-92de-44b7-96b7-59f4238835c7 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1131.118754] env[63372]: INFO nova.compute.manager [req-29ce3d70-56c3-488a-97c6-6f2853d21619 req-01e8a652-031e-4c6c-9d7f-f993280f868a service nova] [instance: c9e5749c-a090-4ce9-b796-852438256502] Neutron deleted interface 4a95ce9d-92de-44b7-96b7-59f4238835c7; detaching it from the instance and deleting it from the info cache [ 1131.118903] env[63372]: DEBUG nova.network.neutron [req-29ce3d70-56c3-488a-97c6-6f2853d21619 req-01e8a652-031e-4c6c-9d7f-f993280f868a service nova] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.340280] env[63372]: DEBUG oslo_concurrency.lockutils [None req-efb54dcc-830b-47df-878a-6001d9a7e183 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.247s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.363386] env[63372]: INFO nova.compute.manager [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Shelve offloading [ 1131.365302] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1131.365555] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d213937-33a0-46da-a612-d6064fcdc9fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.374053] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1131.374053] env[63372]: value = "task-1024885" [ 1131.374053] env[63372]: _type = "Task" [ 1131.374053] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.383742] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] VM already powered off {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1131.383870] env[63372]: DEBUG nova.compute.manager [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1131.385013] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf7e39a-3660-4e9e-b414-cb3656bce882 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.391944] env[63372]: DEBUG oslo_concurrency.lockutils [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.392128] env[63372]: DEBUG oslo_concurrency.lockutils [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.392303] env[63372]: DEBUG nova.network.neutron [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1131.596935] env[63372]: DEBUG nova.network.neutron [-] [instance: c9e5749c-a090-4ce9-b796-852438256502] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.621207] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64d229e2-9cc9-437f-81de-74d2d0953fd3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.630932] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b400a3-b0d7-4953-b155-66cd291345b8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.660644] env[63372]: DEBUG nova.compute.manager [req-29ce3d70-56c3-488a-97c6-6f2853d21619 req-01e8a652-031e-4c6c-9d7f-f993280f868a service nova] [instance: c9e5749c-a090-4ce9-b796-852438256502] Detach interface failed, port_id=4a95ce9d-92de-44b7-96b7-59f4238835c7, reason: Instance c9e5749c-a090-4ce9-b796-852438256502 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1132.099177] env[63372]: INFO nova.compute.manager [-] [instance: c9e5749c-a090-4ce9-b796-852438256502] Took 1.21 seconds to deallocate network for instance. [ 1132.146015] env[63372]: DEBUG nova.network.neutron [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Updating instance_info_cache with network_info: [{"id": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "address": "fa:16:3e:2b:d8:9b", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap099091f6-c2", "ovs_interfaceid": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.579612] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "1e7590de-55ae-4725-be5a-e3435d98151a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.579908] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "1e7590de-55ae-4725-be5a-e3435d98151a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.606098] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.606366] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.606593] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.625944] env[63372]: INFO nova.scheduler.client.report [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Deleted allocations for instance c9e5749c-a090-4ce9-b796-852438256502 [ 1132.647901] env[63372]: DEBUG oslo_concurrency.lockutils [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.005293] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1133.006299] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a8f5346-8235-4118-938c-25f09c796d8e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.014175] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1133.014413] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab76d72e-a297-4d45-a7a5-3385a734e396 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.080199] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1133.080513] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1133.080760] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleting the datastore file [datastore2] 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1133.081085] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-29d529b5-8825-446c-954c-c38b22f5293c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.083551] env[63372]: DEBUG nova.compute.manager [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1133.091877] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1133.091877] env[63372]: value = "task-1024887" [ 1133.091877] env[63372]: _type = "Task" [ 1133.091877] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.100050] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024887, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.134449] env[63372]: DEBUG oslo_concurrency.lockutils [None req-20c5ae75-555b-4042-93cb-17951f77adfb tempest-DeleteServersTestJSON-1365444253 tempest-DeleteServersTestJSON-1365444253-project-member] Lock "c9e5749c-a090-4ce9-b796-852438256502" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.428s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.143510] env[63372]: DEBUG nova.compute.manager [req-a6df0700-f6b3-4869-9e5a-153362662244 req-30fec341-a901-4e0c-a704-3a1add3eecc8 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Received event network-vif-unplugged-099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1133.143674] env[63372]: DEBUG oslo_concurrency.lockutils [req-a6df0700-f6b3-4869-9e5a-153362662244 req-30fec341-a901-4e0c-a704-3a1add3eecc8 service nova] Acquiring lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.143902] env[63372]: DEBUG oslo_concurrency.lockutils [req-a6df0700-f6b3-4869-9e5a-153362662244 req-30fec341-a901-4e0c-a704-3a1add3eecc8 service nova] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.143991] env[63372]: DEBUG oslo_concurrency.lockutils [req-a6df0700-f6b3-4869-9e5a-153362662244 req-30fec341-a901-4e0c-a704-3a1add3eecc8 service nova] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.144172] env[63372]: DEBUG nova.compute.manager [req-a6df0700-f6b3-4869-9e5a-153362662244 req-30fec341-a901-4e0c-a704-3a1add3eecc8 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] No waiting events found dispatching network-vif-unplugged-099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1133.144431] env[63372]: WARNING nova.compute.manager [req-a6df0700-f6b3-4869-9e5a-153362662244 req-30fec341-a901-4e0c-a704-3a1add3eecc8 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Received unexpected event network-vif-unplugged-099091f6-c23b-48d0-9c21-76b4bcc3ad06 for instance with vm_state shelved and task_state shelving_offloading. [ 1133.144509] env[63372]: DEBUG nova.compute.manager [req-a6df0700-f6b3-4869-9e5a-153362662244 req-30fec341-a901-4e0c-a704-3a1add3eecc8 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Received event network-changed-099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1133.144686] env[63372]: DEBUG nova.compute.manager [req-a6df0700-f6b3-4869-9e5a-153362662244 req-30fec341-a901-4e0c-a704-3a1add3eecc8 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Refreshing instance network info cache due to event network-changed-099091f6-c23b-48d0-9c21-76b4bcc3ad06. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1133.144893] env[63372]: DEBUG oslo_concurrency.lockutils [req-a6df0700-f6b3-4869-9e5a-153362662244 req-30fec341-a901-4e0c-a704-3a1add3eecc8 service nova] Acquiring lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.145062] env[63372]: DEBUG oslo_concurrency.lockutils [req-a6df0700-f6b3-4869-9e5a-153362662244 req-30fec341-a901-4e0c-a704-3a1add3eecc8 service nova] Acquired lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.145224] env[63372]: DEBUG nova.network.neutron [req-a6df0700-f6b3-4869-9e5a-153362662244 req-30fec341-a901-4e0c-a704-3a1add3eecc8 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Refreshing network info cache for port 099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1133.601796] env[63372]: DEBUG oslo_vmware.api [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024887, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146846} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.602094] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1133.602285] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1133.602461] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1133.606100] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.606334] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.607831] env[63372]: INFO nova.compute.claims [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1133.621018] env[63372]: INFO nova.scheduler.client.report [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleted allocations for instance 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b [ 1133.863741] env[63372]: DEBUG nova.network.neutron [req-a6df0700-f6b3-4869-9e5a-153362662244 req-30fec341-a901-4e0c-a704-3a1add3eecc8 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Updated VIF entry in instance network info cache for port 099091f6-c23b-48d0-9c21-76b4bcc3ad06. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1133.864172] env[63372]: DEBUG nova.network.neutron [req-a6df0700-f6b3-4869-9e5a-153362662244 req-30fec341-a901-4e0c-a704-3a1add3eecc8 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Updating instance_info_cache with network_info: [{"id": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "address": "fa:16:3e:2b:d8:9b", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": null, "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap099091f6-c2", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.125240] env[63372]: DEBUG oslo_concurrency.lockutils [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.366991] env[63372]: DEBUG oslo_concurrency.lockutils [req-a6df0700-f6b3-4869-9e5a-153362662244 req-30fec341-a901-4e0c-a704-3a1add3eecc8 service nova] Releasing lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.713837] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-573bb8d3-f037-4737-abd3-e0c9c7f0bb95 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.721261] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0fa41b8-b118-4cb6-8303-627eb5d88bc8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.749910] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb65f50-aa5f-47d2-aafe-b9cc6a44ae6c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.756533] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c00a74-ccc8-4dcd-976c-200036bcb861 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.771382] env[63372]: DEBUG nova.compute.provider_tree [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1135.279883] env[63372]: DEBUG nova.scheduler.client.report [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1135.282157] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.783888] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.177s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.784474] env[63372]: DEBUG nova.compute.manager [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1135.787157] env[63372]: DEBUG oslo_concurrency.lockutils [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.662s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.787380] env[63372]: DEBUG nova.objects.instance [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'resources' on Instance uuid 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1136.290296] env[63372]: DEBUG nova.compute.utils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1136.291711] env[63372]: DEBUG nova.compute.manager [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1136.291892] env[63372]: DEBUG nova.network.neutron [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1136.293995] env[63372]: DEBUG nova.objects.instance [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'numa_topology' on Instance uuid 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1136.330832] env[63372]: DEBUG nova.policy [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc5c59e1419843ad90b1bc33f0a55c31', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8452e71bbbd04222bb6b868017b1b6ed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1136.651769] env[63372]: DEBUG nova.network.neutron [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Successfully created port: c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1136.797510] env[63372]: DEBUG nova.compute.manager [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1136.799361] env[63372]: DEBUG nova.objects.base [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Object Instance<2b907d6c-ee2c-4cd5-a77a-babfeb41b88b> lazy-loaded attributes: resources,numa_topology {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1136.919388] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-091daf2b-33a0-4301-8827-7c5cfa95defa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.926951] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634e25e2-633c-4f29-b970-c9cebaa7bbfd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.958047] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0935eee-1cca-4ca8-9e4c-fb1ca8aa1f98 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.965392] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76240e6c-246d-47b2-b11f-4f31fb47c471 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.978588] env[63372]: DEBUG nova.compute.provider_tree [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1137.482228] env[63372]: DEBUG nova.scheduler.client.report [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1137.808858] env[63372]: DEBUG nova.compute.manager [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1137.841078] env[63372]: DEBUG nova.virt.hardware [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1137.841356] env[63372]: DEBUG nova.virt.hardware [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1137.841518] env[63372]: DEBUG nova.virt.hardware [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1137.841731] env[63372]: DEBUG nova.virt.hardware [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1137.841887] env[63372]: DEBUG nova.virt.hardware [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1137.842772] env[63372]: DEBUG nova.virt.hardware [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1137.843127] env[63372]: DEBUG nova.virt.hardware [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1137.843373] env[63372]: DEBUG nova.virt.hardware [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1137.843699] env[63372]: DEBUG nova.virt.hardware [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1137.843945] env[63372]: DEBUG nova.virt.hardware [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1137.844200] env[63372]: DEBUG nova.virt.hardware [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1137.846943] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c170ba-4fd3-4d56-b69e-26d216eea559 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.856907] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b18d865-3e37-4e1d-aac3-2b88b7c5410f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.986818] env[63372]: DEBUG oslo_concurrency.lockutils [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.200s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.212886] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "28eb1675-b1b8-46a6-873d-5c858b716575" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.213196] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "28eb1675-b1b8-46a6-873d-5c858b716575" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.213412] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "28eb1675-b1b8-46a6-873d-5c858b716575-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.213599] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "28eb1675-b1b8-46a6-873d-5c858b716575-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.213774] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "28eb1675-b1b8-46a6-873d-5c858b716575-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.218081] env[63372]: INFO nova.compute.manager [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Terminating instance [ 1138.220353] env[63372]: DEBUG nova.compute.manager [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1138.220756] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1138.222591] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8fdf00-fcd0-4483-9dc1-3c78a30f53ee {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.233359] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1138.233662] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c2a2a7aa-1a16-4e17-bb7f-61739b3fb905 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.251053] env[63372]: DEBUG oslo_vmware.api [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1138.251053] env[63372]: value = "task-1024889" [ 1138.251053] env[63372]: _type = "Task" [ 1138.251053] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.261143] env[63372]: DEBUG oslo_vmware.api [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024889, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.276462] env[63372]: DEBUG nova.compute.manager [req-b55b373a-2ffe-4fb2-88ca-118eb68fcf3b req-0fb3a3a0-c094-49f9-921f-13bb11aae1b6 service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Received event network-vif-plugged-c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1138.276692] env[63372]: DEBUG oslo_concurrency.lockutils [req-b55b373a-2ffe-4fb2-88ca-118eb68fcf3b req-0fb3a3a0-c094-49f9-921f-13bb11aae1b6 service nova] Acquiring lock "1e7590de-55ae-4725-be5a-e3435d98151a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.276965] env[63372]: DEBUG oslo_concurrency.lockutils [req-b55b373a-2ffe-4fb2-88ca-118eb68fcf3b req-0fb3a3a0-c094-49f9-921f-13bb11aae1b6 service nova] Lock "1e7590de-55ae-4725-be5a-e3435d98151a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.277204] env[63372]: DEBUG oslo_concurrency.lockutils [req-b55b373a-2ffe-4fb2-88ca-118eb68fcf3b req-0fb3a3a0-c094-49f9-921f-13bb11aae1b6 service nova] Lock "1e7590de-55ae-4725-be5a-e3435d98151a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.277383] env[63372]: DEBUG nova.compute.manager [req-b55b373a-2ffe-4fb2-88ca-118eb68fcf3b req-0fb3a3a0-c094-49f9-921f-13bb11aae1b6 service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] No waiting events found dispatching network-vif-plugged-c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1138.277552] env[63372]: WARNING nova.compute.manager [req-b55b373a-2ffe-4fb2-88ca-118eb68fcf3b req-0fb3a3a0-c094-49f9-921f-13bb11aae1b6 service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Received unexpected event network-vif-plugged-c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8 for instance with vm_state building and task_state spawning. [ 1138.347053] env[63372]: DEBUG nova.network.neutron [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Successfully updated port: c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1138.496758] env[63372]: DEBUG oslo_concurrency.lockutils [None req-05e084e1-534c-4906-a5f6-26f95f8f1662 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 25.426s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.497471] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 15.226s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.497692] env[63372]: INFO nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] During sync_power_state the instance has a pending task (shelving_image_uploading). Skip. [ 1138.498536] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.498536] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 3.216s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.498633] env[63372]: INFO nova.compute.manager [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Unshelving [ 1138.661154] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.761811] env[63372]: DEBUG oslo_vmware.api [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024889, 'name': PowerOffVM_Task, 'duration_secs': 0.21641} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.761811] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1138.761811] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1138.762062] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed4d00af-347f-4493-ba63-39e9991e1f8c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.853402] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "refresh_cache-1e7590de-55ae-4725-be5a-e3435d98151a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1138.853807] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired lock "refresh_cache-1e7590de-55ae-4725-be5a-e3435d98151a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.853807] env[63372]: DEBUG nova.network.neutron [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1139.124102] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.127303] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.127508] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.127664] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.387268] env[63372]: DEBUG nova.network.neutron [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1139.522321] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.522611] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.522901] env[63372]: DEBUG nova.objects.instance [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'pci_requests' on Instance uuid 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1139.531916] env[63372]: DEBUG nova.network.neutron [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Updating instance_info_cache with network_info: [{"id": "c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8", "address": "fa:16:3e:a5:d5:15", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9c45d69-e9", "ovs_interfaceid": "c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.679920] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1139.680171] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1139.680355] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleting the datastore file [datastore1] 28eb1675-b1b8-46a6-873d-5c858b716575 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1139.680709] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-473aff18-9426-491f-892a-721f568d0893 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.687189] env[63372]: DEBUG oslo_vmware.api [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for the task: (returnval){ [ 1139.687189] env[63372]: value = "task-1024891" [ 1139.687189] env[63372]: _type = "Task" [ 1139.687189] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.695064] env[63372]: DEBUG oslo_vmware.api [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024891, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.027165] env[63372]: DEBUG nova.objects.instance [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'numa_topology' on Instance uuid 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.034368] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Releasing lock "refresh_cache-1e7590de-55ae-4725-be5a-e3435d98151a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.034811] env[63372]: DEBUG nova.compute.manager [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Instance network_info: |[{"id": "c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8", "address": "fa:16:3e:a5:d5:15", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9c45d69-e9", "ovs_interfaceid": "c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1140.035022] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:d5:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1140.042537] env[63372]: DEBUG oslo.service.loopingcall [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1140.042992] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1140.043305] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f08eeb51-d58d-45a1-9796-2b3756bf27a3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.064051] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1140.064051] env[63372]: value = "task-1024892" [ 1140.064051] env[63372]: _type = "Task" [ 1140.064051] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.071768] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024892, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.128351] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1140.128515] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Starting heal instance info cache {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1140.197637] env[63372]: DEBUG oslo_vmware.api [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Task: {'id': task-1024891, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.450946} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.197896] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1140.198097] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1140.198276] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1140.198451] env[63372]: INFO nova.compute.manager [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Took 1.98 seconds to destroy the instance on the hypervisor. [ 1140.198688] env[63372]: DEBUG oslo.service.loopingcall [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1140.198883] env[63372]: DEBUG nova.compute.manager [-] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1140.198994] env[63372]: DEBUG nova.network.neutron [-] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1140.305135] env[63372]: DEBUG nova.compute.manager [req-ecad93e6-94e5-4ead-a837-9598141f9679 req-a21c0229-405f-43cc-adc5-2ada5316aeda service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Received event network-changed-c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1140.305352] env[63372]: DEBUG nova.compute.manager [req-ecad93e6-94e5-4ead-a837-9598141f9679 req-a21c0229-405f-43cc-adc5-2ada5316aeda service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Refreshing instance network info cache due to event network-changed-c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1140.306028] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecad93e6-94e5-4ead-a837-9598141f9679 req-a21c0229-405f-43cc-adc5-2ada5316aeda service nova] Acquiring lock "refresh_cache-1e7590de-55ae-4725-be5a-e3435d98151a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.306202] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecad93e6-94e5-4ead-a837-9598141f9679 req-a21c0229-405f-43cc-adc5-2ada5316aeda service nova] Acquired lock "refresh_cache-1e7590de-55ae-4725-be5a-e3435d98151a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.306373] env[63372]: DEBUG nova.network.neutron [req-ecad93e6-94e5-4ead-a837-9598141f9679 req-a21c0229-405f-43cc-adc5-2ada5316aeda service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Refreshing network info cache for port c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1140.529340] env[63372]: INFO nova.compute.claims [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1140.575179] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024892, 'name': CreateVM_Task, 'duration_secs': 0.341956} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.575354] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1140.576017] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.576191] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.576508] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1140.576758] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8821f2cf-dcdc-4451-b25b-27f0b66dbbf8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.581543] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1140.581543] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52ee0812-0219-6ac7-7f06-588c03afc5a6" [ 1140.581543] env[63372]: _type = "Task" [ 1140.581543] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.589896] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ee0812-0219-6ac7-7f06-588c03afc5a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.660198] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.660343] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquired lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.660490] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Forcefully refreshing network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1141.003050] env[63372]: DEBUG nova.network.neutron [req-ecad93e6-94e5-4ead-a837-9598141f9679 req-a21c0229-405f-43cc-adc5-2ada5316aeda service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Updated VIF entry in instance network info cache for port c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1141.003050] env[63372]: DEBUG nova.network.neutron [req-ecad93e6-94e5-4ead-a837-9598141f9679 req-a21c0229-405f-43cc-adc5-2ada5316aeda service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Updating instance_info_cache with network_info: [{"id": "c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8", "address": "fa:16:3e:a5:d5:15", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9c45d69-e9", "ovs_interfaceid": "c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.092552] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ee0812-0219-6ac7-7f06-588c03afc5a6, 'name': SearchDatastore_Task, 'duration_secs': 0.009685} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.092904] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.093125] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1141.093366] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.093512] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.093692] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1141.093948] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b1d6c39-4db0-49dd-ab5e-4bdb8eebf529 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.101881] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1141.102076] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1141.102815] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb3f0ce0-18bc-4ca1-914b-354acc86b466 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.108147] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1141.108147] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5215fb5a-e3ce-08e1-e586-7e4e9db158ba" [ 1141.108147] env[63372]: _type = "Task" [ 1141.108147] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.115952] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5215fb5a-e3ce-08e1-e586-7e4e9db158ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.148415] env[63372]: DEBUG nova.network.neutron [-] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.506100] env[63372]: DEBUG oslo_concurrency.lockutils [req-ecad93e6-94e5-4ead-a837-9598141f9679 req-a21c0229-405f-43cc-adc5-2ada5316aeda service nova] Releasing lock "refresh_cache-1e7590de-55ae-4725-be5a-e3435d98151a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.619396] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5215fb5a-e3ce-08e1-e586-7e4e9db158ba, 'name': SearchDatastore_Task, 'duration_secs': 0.009006} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.620494] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd0e0ffb-2ba6-453e-ae5b-27ae4bbeeca8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.625464] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1141.625464] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52372f50-57de-a30f-901e-2fdf8bd4388e" [ 1141.625464] env[63372]: _type = "Task" [ 1141.625464] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.635489] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52372f50-57de-a30f-901e-2fdf8bd4388e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.648808] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa701c3-5762-4ef2-baba-805532d1494a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.651434] env[63372]: INFO nova.compute.manager [-] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Took 1.45 seconds to deallocate network for instance. [ 1141.659707] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecea8eac-8abd-4408-a4ee-2d6091105cf3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.691503] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7174894-aa7e-4f81-a6aa-0b210b837d72 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.700896] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5803538-511a-42d8-82f0-ba2522c06c36 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.713942] env[63372]: DEBUG nova.compute.provider_tree [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.861489] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Updating instance_info_cache with network_info: [{"id": "b54d2aaa-a37b-4db7-b735-f9d2a6a9534b", "address": "fa:16:3e:03:b6:cc", "network": {"id": "d47f1dcf-6ba3-42a2-bc84-57879476acf7", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1528873215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ca1f09f7f9e4b2b8010e478202373ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb54d2aaa-a3", "ovs_interfaceid": "b54d2aaa-a37b-4db7-b735-f9d2a6a9534b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.136830] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52372f50-57de-a30f-901e-2fdf8bd4388e, 'name': SearchDatastore_Task, 'duration_secs': 0.008865} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.137165] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.137371] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 1e7590de-55ae-4725-be5a-e3435d98151a/1e7590de-55ae-4725-be5a-e3435d98151a.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1142.137656] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e77d6e3-42e1-4008-83a1-f9f154692eb8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.144139] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1142.144139] env[63372]: value = "task-1024893" [ 1142.144139] env[63372]: _type = "Task" [ 1142.144139] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.152129] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024893, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.158146] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.216598] env[63372]: DEBUG nova.scheduler.client.report [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1142.334452] env[63372]: DEBUG nova.compute.manager [req-78c1bb7a-1a6e-4efb-b46c-f217a4b750a9 req-6094d63a-6fd9-4b27-93a2-d105dba2a560 service nova] [instance: 28eb1675-b1b8-46a6-873d-5c858b716575] Received event network-vif-deleted-845b30fa-3e7e-4324-9479-aed7aade3f91 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1142.363976] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Releasing lock "refresh_cache-ddca7aa5-fcc8-4835-bfdb-47781335f2cd" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.364241] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Updated the network info_cache for instance {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1142.364488] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.364663] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.364834] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63372) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1142.658515] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024893, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.721864] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.199s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.724243] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.566s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.724484] env[63372]: DEBUG nova.objects.instance [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lazy-loading 'resources' on Instance uuid 28eb1675-b1b8-46a6-873d-5c858b716575 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1142.751397] env[63372]: INFO nova.network.neutron [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Updating port 099091f6-c23b-48d0-9c21-76b4bcc3ad06 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1143.156169] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024893, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.695346} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.156579] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] 1e7590de-55ae-4725-be5a-e3435d98151a/1e7590de-55ae-4725-be5a-e3435d98151a.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1143.156693] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1143.156953] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30ece4b1-438e-438d-979e-ada6f3d7f9d5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.166045] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1143.166045] env[63372]: value = "task-1024894" [ 1143.166045] env[63372]: _type = "Task" [ 1143.166045] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.174608] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024894, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.337448] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d96405f-45e9-4c5b-a5fe-8a4959da8988 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.344893] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa1ea945-7657-4ac4-b40b-6166b227c040 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.374648] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9070cc88-9d1f-42f6-8e6f-81f89eb8af66 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.381727] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a57e0e9-45b2-4940-81c9-4deb2ae1cc38 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.394565] env[63372]: DEBUG nova.compute.provider_tree [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.676338] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024894, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113265} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.676954] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1143.677483] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-035b1f6e-514c-4e93-954e-910a949a789c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.701338] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] 1e7590de-55ae-4725-be5a-e3435d98151a/1e7590de-55ae-4725-be5a-e3435d98151a.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1143.701739] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fe1cf4e-ae70-4597-b32c-02e80bf7a745 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.721216] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1143.721216] env[63372]: value = "task-1024895" [ 1143.721216] env[63372]: _type = "Task" [ 1143.721216] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.729182] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024895, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.897912] env[63372]: DEBUG nova.scheduler.client.report [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1144.127930] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager.update_available_resource {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.233294] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024895, 'name': ReconfigVM_Task, 'duration_secs': 0.32848} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.233669] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Reconfigured VM instance instance-00000070 to attach disk [datastore1] 1e7590de-55ae-4725-be5a-e3435d98151a/1e7590de-55ae-4725-be5a-e3435d98151a.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1144.234551] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-90b9f5e1-ba45-4a4c-a605-fe81aaa72f82 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.242939] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1144.242939] env[63372]: value = "task-1024896" [ 1144.242939] env[63372]: _type = "Task" [ 1144.242939] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.252056] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024896, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.294073] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.294073] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.404838] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.678s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.418386] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.418637] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.418790] env[63372]: DEBUG nova.network.neutron [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1144.424671] env[63372]: INFO nova.scheduler.client.report [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Deleted allocations for instance 28eb1675-b1b8-46a6-873d-5c858b716575 [ 1144.434533] env[63372]: DEBUG nova.compute.manager [req-6ea957bb-d097-4665-8a77-bcbd2dd113c4 req-1bdb0427-640b-4d7a-82b1-c6daea0ab81d service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Received event network-vif-plugged-099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1144.434744] env[63372]: DEBUG oslo_concurrency.lockutils [req-6ea957bb-d097-4665-8a77-bcbd2dd113c4 req-1bdb0427-640b-4d7a-82b1-c6daea0ab81d service nova] Acquiring lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.434950] env[63372]: DEBUG oslo_concurrency.lockutils [req-6ea957bb-d097-4665-8a77-bcbd2dd113c4 req-1bdb0427-640b-4d7a-82b1-c6daea0ab81d service nova] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.435150] env[63372]: DEBUG oslo_concurrency.lockutils [req-6ea957bb-d097-4665-8a77-bcbd2dd113c4 req-1bdb0427-640b-4d7a-82b1-c6daea0ab81d service nova] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.435343] env[63372]: DEBUG nova.compute.manager [req-6ea957bb-d097-4665-8a77-bcbd2dd113c4 req-1bdb0427-640b-4d7a-82b1-c6daea0ab81d service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] No waiting events found dispatching network-vif-plugged-099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1144.435480] env[63372]: WARNING nova.compute.manager [req-6ea957bb-d097-4665-8a77-bcbd2dd113c4 req-1bdb0427-640b-4d7a-82b1-c6daea0ab81d service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Received unexpected event network-vif-plugged-099091f6-c23b-48d0-9c21-76b4bcc3ad06 for instance with vm_state shelved_offloaded and task_state spawning. [ 1144.631485] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.631729] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.631874] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.632051] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63372) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1144.632959] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6c7ea4-cf19-411d-9c6b-baaa9260be5e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.641351] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb3a84b2-863e-466b-9702-2f2f8248b3c3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.656182] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751a6f66-229c-4cf6-90cf-dfe6c5031a16 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.662347] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2486663-b9b3-4f32-a51b-887a4876127c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.690718] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180485MB free_disk=185GB free_vcpus=48 pci_devices=None {{(pid=63372) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1144.690862] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.691066] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.751093] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024896, 'name': Rename_Task, 'duration_secs': 0.156618} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.751382] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1144.751616] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70f23e81-4f5e-47b6-92f8-953b18110eaa {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.757894] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1144.757894] env[63372]: value = "task-1024897" [ 1144.757894] env[63372]: _type = "Task" [ 1144.757894] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.764920] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024897, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.796497] env[63372]: INFO nova.compute.manager [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Detaching volume 2b5f42b6-add6-497e-9592-e48a741bd2f0 [ 1144.835344] env[63372]: INFO nova.virt.block_device [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Attempting to driver detach volume 2b5f42b6-add6-497e-9592-e48a741bd2f0 from mountpoint /dev/sdb [ 1144.835782] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Volume detach. Driver type: vmdk {{(pid=63372) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1144.836088] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227463', 'volume_id': '2b5f42b6-add6-497e-9592-e48a741bd2f0', 'name': 'volume-2b5f42b6-add6-497e-9592-e48a741bd2f0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a6bd1c51-e702-48e4-b9ab-8d80db1fec9e', 'attached_at': '', 'detached_at': '', 'volume_id': '2b5f42b6-add6-497e-9592-e48a741bd2f0', 'serial': '2b5f42b6-add6-497e-9592-e48a741bd2f0'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1144.837023] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cbe4375-cbaa-47a3-94f1-1720e20cbae4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.859879] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac0630b-80e1-41e5-aaa3-bc0e1fc236d2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.866964] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba57735-8f77-4ddb-a6c2-66a774870544 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.888492] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c974bf-97f8-4872-b2ba-823347fc159f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.903759] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] The volume has not been displaced from its original location: [datastore1] volume-2b5f42b6-add6-497e-9592-e48a741bd2f0/volume-2b5f42b6-add6-497e-9592-e48a741bd2f0.vmdk. No consolidation needed. {{(pid=63372) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1144.909030] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Reconfiguring VM instance instance-00000068 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1144.909431] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90a20b2c-e1ce-4186-9713-d93fda5cfb76 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.928045] env[63372]: DEBUG oslo_vmware.api [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1144.928045] env[63372]: value = "task-1024898" [ 1144.928045] env[63372]: _type = "Task" [ 1144.928045] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.939516] env[63372]: DEBUG oslo_concurrency.lockutils [None req-70bb7096-7763-4f77-9b08-ce635ce5a0d7 tempest-AttachVolumeShelveTestJSON-208437217 tempest-AttachVolumeShelveTestJSON-208437217-project-member] Lock "28eb1675-b1b8-46a6-873d-5c858b716575" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.726s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.940561] env[63372]: DEBUG oslo_vmware.api [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024898, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.157337] env[63372]: DEBUG nova.network.neutron [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Updating instance_info_cache with network_info: [{"id": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "address": "fa:16:3e:2b:d8:9b", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap099091f6-c2", "ovs_interfaceid": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.267796] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024897, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.437393] env[63372]: DEBUG oslo_vmware.api [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024898, 'name': ReconfigVM_Task, 'duration_secs': 0.228113} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.437674] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Reconfigured VM instance instance-00000068 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1145.442589] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d72c726-e2fe-4c14-8f3a-5c4d7e80aa14 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.457871] env[63372]: DEBUG oslo_vmware.api [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1145.457871] env[63372]: value = "task-1024899" [ 1145.457871] env[63372]: _type = "Task" [ 1145.457871] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.468287] env[63372]: DEBUG oslo_vmware.api [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024899, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.660477] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.687478] env[63372]: DEBUG nova.virt.hardware [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='0015c945c9894092e0bfa07a297daea3',container_format='bare',created_at=2024-09-30T11:44:29Z,direct_url=,disk_format='vmdk',id=53b2f81e-8df7-4596-909e-8ed3e82311c7,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-2127688635-shelved',owner='98581685387a4f1499ae6ed378af982c',properties=ImageMetaProps,protected=,size=31666688,status='active',tags=,updated_at=2024-09-30T11:44:46Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1145.687730] env[63372]: DEBUG nova.virt.hardware [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1145.687889] env[63372]: DEBUG nova.virt.hardware [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1145.688087] env[63372]: DEBUG nova.virt.hardware [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1145.688240] env[63372]: DEBUG nova.virt.hardware [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1145.688387] env[63372]: DEBUG nova.virt.hardware [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1145.688592] env[63372]: DEBUG nova.virt.hardware [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1145.688755] env[63372]: DEBUG nova.virt.hardware [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1145.688921] env[63372]: DEBUG nova.virt.hardware [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1145.689151] env[63372]: DEBUG nova.virt.hardware [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1145.689284] env[63372]: DEBUG nova.virt.hardware [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1145.690236] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f95f7b72-cc28-4206-8235-6a91fe02f87a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.698281] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79abe346-f540-482e-b8cc-5e753adf172c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.714617] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:d8:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '099091f6-c23b-48d0-9c21-76b4bcc3ad06', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1145.721979] env[63372]: DEBUG oslo.service.loopingcall [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1145.722858] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance ddca7aa5-fcc8-4835-bfdb-47781335f2cd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.722995] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 1724ce03-c8e2-415d-a380-59ac69fbfb57 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.723128] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance a6bd1c51-e702-48e4-b9ab-8d80db1fec9e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.723242] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.723354] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 802deb8b-978b-4d01-b19c-a0090e0029f0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.723463] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance fafed45e-c9dc-4196-883f-a23570fd0583 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.723571] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 1e7590de-55ae-4725-be5a-e3435d98151a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.723681] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.723857] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1145.723987] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1145.725929] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1145.726693] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc08e057-3979-44a6-b5a0-57d0c5f348ba {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.747974] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1145.747974] env[63372]: value = "task-1024900" [ 1145.747974] env[63372]: _type = "Task" [ 1145.747974] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.758915] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024900, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.768331] env[63372]: DEBUG oslo_vmware.api [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024897, 'name': PowerOnVM_Task, 'duration_secs': 0.534976} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.768579] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1145.768784] env[63372]: INFO nova.compute.manager [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Took 7.96 seconds to spawn the instance on the hypervisor. [ 1145.768959] env[63372]: DEBUG nova.compute.manager [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1145.769674] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a56dc8f-660b-438b-a0f7-935bb2342d33 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.826170] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d858f42b-de01-444e-8e20-300d04046afe {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.834407] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90f9f3f-92e8-406d-99df-d20c9c411160 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.865690] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6880eb-1eb2-4949-80a0-050835c7c76c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.873180] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5a856e-38c8-4edf-82f2-3fb93a2b38b6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.886354] env[63372]: DEBUG nova.compute.provider_tree [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1145.970173] env[63372]: DEBUG oslo_vmware.api [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024899, 'name': ReconfigVM_Task, 'duration_secs': 0.13185} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.970173] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227463', 'volume_id': '2b5f42b6-add6-497e-9592-e48a741bd2f0', 'name': 'volume-2b5f42b6-add6-497e-9592-e48a741bd2f0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a6bd1c51-e702-48e4-b9ab-8d80db1fec9e', 'attached_at': '', 'detached_at': '', 'volume_id': '2b5f42b6-add6-497e-9592-e48a741bd2f0', 'serial': '2b5f42b6-add6-497e-9592-e48a741bd2f0'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1146.257822] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024900, 'name': CreateVM_Task, 'duration_secs': 0.324632} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.257822] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1146.258529] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/53b2f81e-8df7-4596-909e-8ed3e82311c7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.258722] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "[datastore2] devstack-image-cache_base/53b2f81e-8df7-4596-909e-8ed3e82311c7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.259132] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/53b2f81e-8df7-4596-909e-8ed3e82311c7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1146.259396] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b1c8a75-5cf2-4542-89c6-b8b070e2214e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.264179] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1146.264179] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]525c5bf3-6ad9-5e28-1e90-f695342fee71" [ 1146.264179] env[63372]: _type = "Task" [ 1146.264179] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.271273] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]525c5bf3-6ad9-5e28-1e90-f695342fee71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.293312] env[63372]: INFO nova.compute.manager [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Took 12.70 seconds to build instance. [ 1146.389131] env[63372]: DEBUG nova.scheduler.client.report [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1146.463037] env[63372]: DEBUG nova.compute.manager [req-27598374-acd8-4445-b0e8-3bb3c6485e50 req-27eb4779-cef6-42c3-864d-5c33cd0dec55 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Received event network-changed-099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1146.463264] env[63372]: DEBUG nova.compute.manager [req-27598374-acd8-4445-b0e8-3bb3c6485e50 req-27eb4779-cef6-42c3-864d-5c33cd0dec55 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Refreshing instance network info cache due to event network-changed-099091f6-c23b-48d0-9c21-76b4bcc3ad06. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1146.463482] env[63372]: DEBUG oslo_concurrency.lockutils [req-27598374-acd8-4445-b0e8-3bb3c6485e50 req-27eb4779-cef6-42c3-864d-5c33cd0dec55 service nova] Acquiring lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.463626] env[63372]: DEBUG oslo_concurrency.lockutils [req-27598374-acd8-4445-b0e8-3bb3c6485e50 req-27eb4779-cef6-42c3-864d-5c33cd0dec55 service nova] Acquired lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.463831] env[63372]: DEBUG nova.network.neutron [req-27598374-acd8-4445-b0e8-3bb3c6485e50 req-27eb4779-cef6-42c3-864d-5c33cd0dec55 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Refreshing network info cache for port 099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1146.512115] env[63372]: DEBUG nova.objects.instance [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lazy-loading 'flavor' on Instance uuid a6bd1c51-e702-48e4-b9ab-8d80db1fec9e {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1146.774710] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "[datastore2] devstack-image-cache_base/53b2f81e-8df7-4596-909e-8ed3e82311c7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.774963] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Processing image 53b2f81e-8df7-4596-909e-8ed3e82311c7 {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1146.775221] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/53b2f81e-8df7-4596-909e-8ed3e82311c7/53b2f81e-8df7-4596-909e-8ed3e82311c7.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.775370] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquired lock "[datastore2] devstack-image-cache_base/53b2f81e-8df7-4596-909e-8ed3e82311c7/53b2f81e-8df7-4596-909e-8ed3e82311c7.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.775549] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1146.775785] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-762c36a2-57ea-4fbd-83f3-9bc6c0a5ef1d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.783531] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1146.783698] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1146.784683] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6614ee77-5bfd-4ec5-8586-8789dd33d6bf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.789357] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1146.789357] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521a5062-8b16-d5d0-3349-1fe29a1405c0" [ 1146.789357] env[63372]: _type = "Task" [ 1146.789357] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.796351] env[63372]: DEBUG oslo_concurrency.lockutils [None req-2f5b3a44-b4db-4155-85c6-2e967af93f6f tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "1e7590de-55ae-4725-be5a-e3435d98151a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.216s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.796577] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521a5062-8b16-d5d0-3349-1fe29a1405c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.866185] env[63372]: DEBUG nova.compute.manager [req-fff02cf5-33e7-4d2f-97aa-4ad6cf4bb2b5 req-50d62c52-99bc-461f-a59b-bef6a19d10f4 service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Received event network-changed-c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1146.866185] env[63372]: DEBUG nova.compute.manager [req-fff02cf5-33e7-4d2f-97aa-4ad6cf4bb2b5 req-50d62c52-99bc-461f-a59b-bef6a19d10f4 service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Refreshing instance network info cache due to event network-changed-c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1146.866185] env[63372]: DEBUG oslo_concurrency.lockutils [req-fff02cf5-33e7-4d2f-97aa-4ad6cf4bb2b5 req-50d62c52-99bc-461f-a59b-bef6a19d10f4 service nova] Acquiring lock "refresh_cache-1e7590de-55ae-4725-be5a-e3435d98151a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.866185] env[63372]: DEBUG oslo_concurrency.lockutils [req-fff02cf5-33e7-4d2f-97aa-4ad6cf4bb2b5 req-50d62c52-99bc-461f-a59b-bef6a19d10f4 service nova] Acquired lock "refresh_cache-1e7590de-55ae-4725-be5a-e3435d98151a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.866185] env[63372]: DEBUG nova.network.neutron [req-fff02cf5-33e7-4d2f-97aa-4ad6cf4bb2b5 req-50d62c52-99bc-461f-a59b-bef6a19d10f4 service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Refreshing network info cache for port c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1146.894287] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1146.894487] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.203s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.212622] env[63372]: DEBUG nova.network.neutron [req-27598374-acd8-4445-b0e8-3bb3c6485e50 req-27eb4779-cef6-42c3-864d-5c33cd0dec55 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Updated VIF entry in instance network info cache for port 099091f6-c23b-48d0-9c21-76b4bcc3ad06. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1147.213010] env[63372]: DEBUG nova.network.neutron [req-27598374-acd8-4445-b0e8-3bb3c6485e50 req-27eb4779-cef6-42c3-864d-5c33cd0dec55 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Updating instance_info_cache with network_info: [{"id": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "address": "fa:16:3e:2b:d8:9b", "network": {"id": "7a447338-1c5b-490b-9b09-17f685d5c710", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-970578692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98581685387a4f1499ae6ed378af982c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap099091f6-c2", "ovs_interfaceid": "099091f6-c23b-48d0-9c21-76b4bcc3ad06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.300045] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Preparing fetch location {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1147.300400] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Fetch image to [datastore2] OSTACK_IMG_da53599c-74e4-4a4c-8400-37166131ac23/OSTACK_IMG_da53599c-74e4-4a4c-8400-37166131ac23.vmdk {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1147.300400] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Downloading stream optimized image 53b2f81e-8df7-4596-909e-8ed3e82311c7 to [datastore2] OSTACK_IMG_da53599c-74e4-4a4c-8400-37166131ac23/OSTACK_IMG_da53599c-74e4-4a4c-8400-37166131ac23.vmdk on the data store datastore2 as vApp {{(pid=63372) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1147.300574] env[63372]: DEBUG nova.virt.vmwareapi.images [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Downloading image file data 53b2f81e-8df7-4596-909e-8ed3e82311c7 to the ESX as VM named 'OSTACK_IMG_da53599c-74e4-4a4c-8400-37166131ac23' {{(pid=63372) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1147.373720] env[63372]: DEBUG oslo_vmware.rw_handles [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1147.373720] env[63372]: value = "resgroup-9" [ 1147.373720] env[63372]: _type = "ResourcePool" [ 1147.373720] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1147.374167] env[63372]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-10dd9e51-b0b3-409f-b2df-60f47b248f16 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.397973] env[63372]: DEBUG oslo_vmware.rw_handles [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lease: (returnval){ [ 1147.397973] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5286a14d-ce05-e098-fa18-1ca0b8952e69" [ 1147.397973] env[63372]: _type = "HttpNfcLease" [ 1147.397973] env[63372]: } obtained for vApp import into resource pool (val){ [ 1147.397973] env[63372]: value = "resgroup-9" [ 1147.397973] env[63372]: _type = "ResourcePool" [ 1147.397973] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1147.398319] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the lease: (returnval){ [ 1147.398319] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5286a14d-ce05-e098-fa18-1ca0b8952e69" [ 1147.398319] env[63372]: _type = "HttpNfcLease" [ 1147.398319] env[63372]: } to be ready. {{(pid=63372) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1147.407385] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1147.407385] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5286a14d-ce05-e098-fa18-1ca0b8952e69" [ 1147.407385] env[63372]: _type = "HttpNfcLease" [ 1147.407385] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1147.518658] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a6ffec4e-8e47-4e3b-bac1-914de13b8eaa tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.225s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.593932] env[63372]: DEBUG nova.network.neutron [req-fff02cf5-33e7-4d2f-97aa-4ad6cf4bb2b5 req-50d62c52-99bc-461f-a59b-bef6a19d10f4 service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Updated VIF entry in instance network info cache for port c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1147.594433] env[63372]: DEBUG nova.network.neutron [req-fff02cf5-33e7-4d2f-97aa-4ad6cf4bb2b5 req-50d62c52-99bc-461f-a59b-bef6a19d10f4 service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Updating instance_info_cache with network_info: [{"id": "c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8", "address": "fa:16:3e:a5:d5:15", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9c45d69-e9", "ovs_interfaceid": "c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.715803] env[63372]: DEBUG oslo_concurrency.lockutils [req-27598374-acd8-4445-b0e8-3bb3c6485e50 req-27eb4779-cef6-42c3-864d-5c33cd0dec55 service nova] Releasing lock "refresh_cache-2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.911540] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1147.911540] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5286a14d-ce05-e098-fa18-1ca0b8952e69" [ 1147.911540] env[63372]: _type = "HttpNfcLease" [ 1147.911540] env[63372]: } is initializing. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1148.097130] env[63372]: DEBUG oslo_concurrency.lockutils [req-fff02cf5-33e7-4d2f-97aa-4ad6cf4bb2b5 req-50d62c52-99bc-461f-a59b-bef6a19d10f4 service nova] Releasing lock "refresh_cache-1e7590de-55ae-4725-be5a-e3435d98151a" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.410271] env[63372]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1148.410271] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5286a14d-ce05-e098-fa18-1ca0b8952e69" [ 1148.410271] env[63372]: _type = "HttpNfcLease" [ 1148.410271] env[63372]: } is ready. {{(pid=63372) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1148.410726] env[63372]: DEBUG oslo_vmware.rw_handles [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1148.410726] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5286a14d-ce05-e098-fa18-1ca0b8952e69" [ 1148.410726] env[63372]: _type = "HttpNfcLease" [ 1148.410726] env[63372]: }. {{(pid=63372) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1148.411396] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ece13ab-98db-431d-aa5f-c8a99e75b73a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.418574] env[63372]: DEBUG oslo_vmware.rw_handles [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e43a84-9c58-9991-98e2-a2b53e319c76/disk-0.vmdk from lease info. {{(pid=63372) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1148.418758] env[63372]: DEBUG oslo_vmware.rw_handles [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Creating HTTP connection to write to file with size = 31666688 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e43a84-9c58-9991-98e2-a2b53e319c76/disk-0.vmdk. {{(pid=63372) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1148.482323] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f5eb5325-5ea9-4b50-8912-79c889ff03e6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.624423] env[63372]: DEBUG oslo_concurrency.lockutils [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.624723] env[63372]: DEBUG oslo_concurrency.lockutils [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.624943] env[63372]: DEBUG oslo_concurrency.lockutils [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.625162] env[63372]: DEBUG oslo_concurrency.lockutils [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.625343] env[63372]: DEBUG oslo_concurrency.lockutils [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.627769] env[63372]: INFO nova.compute.manager [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Terminating instance [ 1148.629609] env[63372]: DEBUG nova.compute.manager [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1148.629819] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1148.630704] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af7e03c0-debf-4024-9631-e3a39768911a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.638701] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1148.638925] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-340e8334-8641-43ae-9398-9329ca4541ed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.645247] env[63372]: DEBUG oslo_vmware.api [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1148.645247] env[63372]: value = "task-1024903" [ 1148.645247] env[63372]: _type = "Task" [ 1148.645247] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.653674] env[63372]: DEBUG oslo_vmware.api [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024903, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.144745] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d9cc789-a136-4fbd-bed2-3afcbba7e847 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "802deb8b-978b-4d01-b19c-a0090e0029f0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.145101] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d9cc789-a136-4fbd-bed2-3afcbba7e847 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "802deb8b-978b-4d01-b19c-a0090e0029f0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.145359] env[63372]: DEBUG nova.compute.manager [None req-3d9cc789-a136-4fbd-bed2-3afcbba7e847 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1149.146432] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58025604-b08d-4bca-96c5-35795cc1c7cb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.160342] env[63372]: DEBUG nova.compute.manager [None req-3d9cc789-a136-4fbd-bed2-3afcbba7e847 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63372) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1149.161074] env[63372]: DEBUG nova.objects.instance [None req-3d9cc789-a136-4fbd-bed2-3afcbba7e847 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lazy-loading 'flavor' on Instance uuid 802deb8b-978b-4d01-b19c-a0090e0029f0 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1149.167422] env[63372]: DEBUG oslo_vmware.api [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024903, 'name': PowerOffVM_Task, 'duration_secs': 0.223579} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.167668] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1149.167840] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1149.168193] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9908bc26-cc08-4b27-8623-5a35d41a5bbd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.242135] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1149.242366] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1149.242561] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Deleting the datastore file [datastore1] a6bd1c51-e702-48e4-b9ab-8d80db1fec9e {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1149.244185] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0cbddb3f-c364-40f5-927d-2236e154956e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.252164] env[63372]: DEBUG oslo_vmware.api [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1149.252164] env[63372]: value = "task-1024905" [ 1149.252164] env[63372]: _type = "Task" [ 1149.252164] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.262037] env[63372]: DEBUG oslo_vmware.api [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024905, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.666328] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d9cc789-a136-4fbd-bed2-3afcbba7e847 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1149.666682] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48e4e242-45cd-41cb-8e6d-ef5d1e17846b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.673596] env[63372]: DEBUG oslo_vmware.api [None req-3d9cc789-a136-4fbd-bed2-3afcbba7e847 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1149.673596] env[63372]: value = "task-1024906" [ 1149.673596] env[63372]: _type = "Task" [ 1149.673596] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.682335] env[63372]: DEBUG oslo_vmware.api [None req-3d9cc789-a136-4fbd-bed2-3afcbba7e847 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024906, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.764032] env[63372]: DEBUG oslo_vmware.api [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024905, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.255605} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.765772] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1149.766054] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1149.766222] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1149.766404] env[63372]: INFO nova.compute.manager [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1149.766740] env[63372]: DEBUG oslo.service.loopingcall [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1149.766995] env[63372]: DEBUG nova.compute.manager [-] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1149.767159] env[63372]: DEBUG nova.network.neutron [-] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1149.782697] env[63372]: DEBUG oslo_vmware.rw_handles [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Completed reading data from the image iterator. {{(pid=63372) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1149.782985] env[63372]: DEBUG oslo_vmware.rw_handles [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e43a84-9c58-9991-98e2-a2b53e319c76/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1149.783929] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32738db7-723b-4f7f-a015-97e9e344c6fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.790676] env[63372]: DEBUG oslo_vmware.rw_handles [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e43a84-9c58-9991-98e2-a2b53e319c76/disk-0.vmdk is in state: ready. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1149.790867] env[63372]: DEBUG oslo_vmware.rw_handles [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e43a84-9c58-9991-98e2-a2b53e319c76/disk-0.vmdk. {{(pid=63372) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1149.791128] env[63372]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-14542006-5036-4b96-8eed-3e900e8277fb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.180883] env[63372]: DEBUG oslo_vmware.rw_handles [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e43a84-9c58-9991-98e2-a2b53e319c76/disk-0.vmdk. {{(pid=63372) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1150.181112] env[63372]: INFO nova.virt.vmwareapi.images [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Downloaded image file data 53b2f81e-8df7-4596-909e-8ed3e82311c7 [ 1150.182317] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a42289-1380-4236-bb33-762ce3f8b52d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.187527] env[63372]: DEBUG oslo_vmware.api [None req-3d9cc789-a136-4fbd-bed2-3afcbba7e847 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024906, 'name': PowerOffVM_Task, 'duration_secs': 0.218447} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.188122] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d9cc789-a136-4fbd-bed2-3afcbba7e847 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1150.188329] env[63372]: DEBUG nova.compute.manager [None req-3d9cc789-a136-4fbd-bed2-3afcbba7e847 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1150.189059] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50a7b55-a094-482a-8761-952199ffbce4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.202098] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5bd07b23-9c1f-4f9a-b5c4-92140effe937 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.418985] env[63372]: INFO nova.virt.vmwareapi.images [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] The imported VM was unregistered [ 1150.424718] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Caching image {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1150.424959] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Creating directory with path [datastore2] devstack-image-cache_base/53b2f81e-8df7-4596-909e-8ed3e82311c7 {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1150.425269] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b0f0d8b-7b74-4991-913d-4d3514fbef82 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.433659] env[63372]: DEBUG nova.compute.manager [req-e6777f92-1147-4a5b-b31c-f2b6cb4edc9d req-e0b0beac-55aa-4fe7-b5f5-097568bffc39 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Received event network-vif-deleted-0fbd29da-197e-452a-9644-ab461d995ec3 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1150.433831] env[63372]: INFO nova.compute.manager [req-e6777f92-1147-4a5b-b31c-f2b6cb4edc9d req-e0b0beac-55aa-4fe7-b5f5-097568bffc39 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Neutron deleted interface 0fbd29da-197e-452a-9644-ab461d995ec3; detaching it from the instance and deleting it from the info cache [ 1150.433966] env[63372]: DEBUG nova.network.neutron [req-e6777f92-1147-4a5b-b31c-f2b6cb4edc9d req-e0b0beac-55aa-4fe7-b5f5-097568bffc39 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1150.442317] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Created directory with path [datastore2] devstack-image-cache_base/53b2f81e-8df7-4596-909e-8ed3e82311c7 {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1150.442667] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_da53599c-74e4-4a4c-8400-37166131ac23/OSTACK_IMG_da53599c-74e4-4a4c-8400-37166131ac23.vmdk to [datastore2] devstack-image-cache_base/53b2f81e-8df7-4596-909e-8ed3e82311c7/53b2f81e-8df7-4596-909e-8ed3e82311c7.vmdk. {{(pid=63372) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1150.443570] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-a147c8d3-6c6e-451f-96d7-092a0fdeef3e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.453963] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1150.453963] env[63372]: value = "task-1024908" [ 1150.453963] env[63372]: _type = "Task" [ 1150.453963] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.464467] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024908, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.711830] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3d9cc789-a136-4fbd-bed2-3afcbba7e847 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "802deb8b-978b-4d01-b19c-a0090e0029f0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.567s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.910414] env[63372]: DEBUG nova.network.neutron [-] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1150.944829] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c41fec0-b62b-4a14-9e81-e7b6502a51c4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.963840] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af4548a-63e1-473f-9ef8-99598dc813df {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.984746] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024908, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.995193] env[63372]: DEBUG nova.compute.manager [req-e6777f92-1147-4a5b-b31c-f2b6cb4edc9d req-e0b0beac-55aa-4fe7-b5f5-097568bffc39 service nova] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Detach interface failed, port_id=0fbd29da-197e-452a-9644-ab461d995ec3, reason: Instance a6bd1c51-e702-48e4-b9ab-8d80db1fec9e could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1151.413803] env[63372]: INFO nova.compute.manager [-] [instance: a6bd1c51-e702-48e4-b9ab-8d80db1fec9e] Took 1.65 seconds to deallocate network for instance. [ 1151.475819] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024908, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.644269] env[63372]: DEBUG nova.objects.instance [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lazy-loading 'flavor' on Instance uuid 802deb8b-978b-4d01-b19c-a0090e0029f0 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1151.923407] env[63372]: DEBUG oslo_concurrency.lockutils [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.924741] env[63372]: DEBUG oslo_concurrency.lockutils [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.924741] env[63372]: DEBUG nova.objects.instance [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lazy-loading 'resources' on Instance uuid a6bd1c51-e702-48e4-b9ab-8d80db1fec9e {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1151.966706] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024908, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.150152] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.150363] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.151639] env[63372]: DEBUG nova.network.neutron [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1152.151639] env[63372]: DEBUG nova.objects.instance [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lazy-loading 'info_cache' on Instance uuid 802deb8b-978b-4d01-b19c-a0090e0029f0 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1152.467208] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024908, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.550821] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a34ebb23-724b-4b15-8c47-89e4dbba2b57 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.559279] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4186a3-82f2-445d-98d5-265d47bf86d1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.590398] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8361af-3501-4122-b490-2f7836567538 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.598380] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b3d403-d496-4d24-9980-507b66b42a91 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.614032] env[63372]: DEBUG nova.compute.provider_tree [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1152.654430] env[63372]: DEBUG nova.objects.base [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Object Instance<802deb8b-978b-4d01-b19c-a0090e0029f0> lazy-loaded attributes: flavor,info_cache {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1152.966801] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024908, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.283682} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.967171] env[63372]: INFO nova.virt.vmwareapi.ds_util [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_da53599c-74e4-4a4c-8400-37166131ac23/OSTACK_IMG_da53599c-74e4-4a4c-8400-37166131ac23.vmdk to [datastore2] devstack-image-cache_base/53b2f81e-8df7-4596-909e-8ed3e82311c7/53b2f81e-8df7-4596-909e-8ed3e82311c7.vmdk. [ 1152.967302] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Cleaning up location [datastore2] OSTACK_IMG_da53599c-74e4-4a4c-8400-37166131ac23 {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1152.967445] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_da53599c-74e4-4a4c-8400-37166131ac23 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1152.967696] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8714db1b-c890-4c9f-8fdc-fee77cb32a44 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.973852] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1152.973852] env[63372]: value = "task-1024909" [ 1152.973852] env[63372]: _type = "Task" [ 1152.973852] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.981557] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024909, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.117131] env[63372]: DEBUG nova.scheduler.client.report [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1153.357521] env[63372]: DEBUG nova.network.neutron [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Updating instance_info_cache with network_info: [{"id": "146031c4-2f32-4085-9fc6-3060e7db9d08", "address": "fa:16:3e:70:3b:d5", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap146031c4-2f", "ovs_interfaceid": "146031c4-2f32-4085-9fc6-3060e7db9d08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.483999] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024909, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.039105} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.484292] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1153.484466] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Releasing lock "[datastore2] devstack-image-cache_base/53b2f81e-8df7-4596-909e-8ed3e82311c7/53b2f81e-8df7-4596-909e-8ed3e82311c7.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.484729] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/53b2f81e-8df7-4596-909e-8ed3e82311c7/53b2f81e-8df7-4596-909e-8ed3e82311c7.vmdk to [datastore2] 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b/2b907d6c-ee2c-4cd5-a77a-babfeb41b88b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1153.484982] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-23ffb1f7-3264-4204-ad61-06476fc89ac0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.491025] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1153.491025] env[63372]: value = "task-1024910" [ 1153.491025] env[63372]: _type = "Task" [ 1153.491025] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.498632] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024910, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.623032] env[63372]: DEBUG oslo_concurrency.lockutils [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.699s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.650534] env[63372]: INFO nova.scheduler.client.report [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Deleted allocations for instance a6bd1c51-e702-48e4-b9ab-8d80db1fec9e [ 1153.860750] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.003078] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024910, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.158479] env[63372]: DEBUG oslo_concurrency.lockutils [None req-72ad4c7f-9e5f-4d5b-ac22-9d16b46da8d7 tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "a6bd1c51-e702-48e4-b9ab-8d80db1fec9e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.534s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.364644] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1154.365031] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-076b86ad-3e72-4753-9272-d92359e79de4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.373007] env[63372]: DEBUG oslo_vmware.api [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1154.373007] env[63372]: value = "task-1024911" [ 1154.373007] env[63372]: _type = "Task" [ 1154.373007] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.381279] env[63372]: DEBUG oslo_vmware.api [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024911, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.502073] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024910, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.885221] env[63372]: DEBUG oslo_vmware.api [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024911, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.002114] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024910, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.365652] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "fafed45e-c9dc-4196-883f-a23570fd0583" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.365984] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.383918] env[63372]: DEBUG oslo_vmware.api [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024911, 'name': PowerOnVM_Task, 'duration_secs': 0.591429} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.384170] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1155.384365] env[63372]: DEBUG nova.compute.manager [None req-b4ef1efb-ea08-4521-a381-675f119361a3 tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1155.385249] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119b13b1-aff1-4531-baff-fe5d355584a0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.502613] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024910, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.869519] env[63372]: DEBUG nova.compute.utils [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1156.003104] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024910, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.3042} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.003349] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/53b2f81e-8df7-4596-909e-8ed3e82311c7/53b2f81e-8df7-4596-909e-8ed3e82311c7.vmdk to [datastore2] 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b/2b907d6c-ee2c-4cd5-a77a-babfeb41b88b.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1156.004157] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a22e3e2-d078-4646-adce-fdcca6ce67dd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.025836] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b/2b907d6c-ee2c-4cd5-a77a-babfeb41b88b.vmdk or device None with type streamOptimized {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1156.026121] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de41e14e-4ecf-4b04-a63b-a82de975028f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.045021] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1156.045021] env[63372]: value = "task-1024913" [ 1156.045021] env[63372]: _type = "Task" [ 1156.045021] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.052567] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024913, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.372701] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.554724] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024913, 'name': ReconfigVM_Task, 'duration_secs': 0.29266} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.555842] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b/2b907d6c-ee2c-4cd5-a77a-babfeb41b88b.vmdk or device None with type streamOptimized {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1156.555981] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-46c86158-4ab2-489a-930d-07d259a6d384 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.562860] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1156.562860] env[63372]: value = "task-1024914" [ 1156.562860] env[63372]: _type = "Task" [ 1156.562860] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.571994] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024914, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.924994] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cfb4a89-ef3d-45d4-b4ac-4034c3996d72 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.935835] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-785bd5d7-ff98-4abd-92ad-badc008d985b tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Suspending the VM {{(pid=63372) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1156.936159] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-ba663673-3249-4b79-9b33-261c631b6341 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.945210] env[63372]: DEBUG oslo_vmware.api [None req-785bd5d7-ff98-4abd-92ad-badc008d985b tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1156.945210] env[63372]: value = "task-1024915" [ 1156.945210] env[63372]: _type = "Task" [ 1156.945210] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.959485] env[63372]: DEBUG oslo_vmware.api [None req-785bd5d7-ff98-4abd-92ad-badc008d985b tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024915, 'name': SuspendVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.073292] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024914, 'name': Rename_Task, 'duration_secs': 0.139475} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.073587] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1157.074166] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09bee217-8278-42fa-a568-8fc20def1587 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.080459] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1157.080459] env[63372]: value = "task-1024916" [ 1157.080459] env[63372]: _type = "Task" [ 1157.080459] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.094220] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024916, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.433080] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "fafed45e-c9dc-4196-883f-a23570fd0583" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.433451] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.433590] env[63372]: INFO nova.compute.manager [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Attaching volume 598ca7c3-f9fb-4fe4-939a-9ca37fc19323 to /dev/sdb [ 1157.454811] env[63372]: DEBUG oslo_vmware.api [None req-785bd5d7-ff98-4abd-92ad-badc008d985b tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024915, 'name': SuspendVM_Task} progress is 62%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.466641] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eaa1e09-2a95-4612-b872-7c38f438d321 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.472673] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8dbcdf-600c-48cc-9df2-89908abcfe60 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.486313] env[63372]: DEBUG nova.virt.block_device [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Updating existing volume attachment record: bc5e437a-14ca-44cf-860a-a43a7193bd45 {{(pid=63372) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1157.590738] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024916, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.955468] env[63372]: DEBUG oslo_vmware.api [None req-785bd5d7-ff98-4abd-92ad-badc008d985b tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024915, 'name': SuspendVM_Task, 'duration_secs': 0.626029} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.955978] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-785bd5d7-ff98-4abd-92ad-badc008d985b tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Suspended the VM {{(pid=63372) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1157.957071] env[63372]: DEBUG nova.compute.manager [None req-785bd5d7-ff98-4abd-92ad-badc008d985b tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1157.957234] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3dcb69c-912a-4a9f-b4e1-81f72fb2af5d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.090134] env[63372]: DEBUG oslo_vmware.api [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024916, 'name': PowerOnVM_Task, 'duration_secs': 0.517927} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.090412] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1158.186392] env[63372]: DEBUG nova.compute.manager [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1158.187315] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7c3025-8b1a-4188-b1c6-03fab0eb0f66 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.705846] env[63372]: DEBUG oslo_concurrency.lockutils [None req-badb2fac-c68b-4dc7-80fc-ae2fb753fff1 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.207s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.304070] env[63372]: INFO nova.compute.manager [None req-704d420c-b279-441f-a95c-1a1d114d8cfc tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Resuming [ 1159.304283] env[63372]: DEBUG nova.objects.instance [None req-704d420c-b279-441f-a95c-1a1d114d8cfc tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lazy-loading 'flavor' on Instance uuid 802deb8b-978b-4d01-b19c-a0090e0029f0 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1159.890094] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.890456] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.890531] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.890720] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.890897] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.892920] env[63372]: INFO nova.compute.manager [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Terminating instance [ 1159.894604] env[63372]: DEBUG nova.compute.manager [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1159.894796] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1159.895623] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f48cbb-e0b8-4e34-b9c1-12fe01241a0c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.903471] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1159.903807] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "1724ce03-c8e2-415d-a380-59ac69fbfb57" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.904025] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "1724ce03-c8e2-415d-a380-59ac69fbfb57" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.904217] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "1724ce03-c8e2-415d-a380-59ac69fbfb57-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.904400] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "1724ce03-c8e2-415d-a380-59ac69fbfb57-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.904567] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "1724ce03-c8e2-415d-a380-59ac69fbfb57-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.905841] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a0e20987-bf37-4c2b-a99d-9ac686574607 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.907555] env[63372]: INFO nova.compute.manager [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Terminating instance [ 1159.909423] env[63372]: DEBUG nova.compute.manager [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1159.909615] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1159.910346] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81180e4c-3ac6-4e33-901d-a0bab0942ac9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.914010] env[63372]: DEBUG oslo_vmware.api [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1159.914010] env[63372]: value = "task-1024920" [ 1159.914010] env[63372]: _type = "Task" [ 1159.914010] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.920596] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1159.921413] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72f74f7c-f933-464a-aa2f-8b9c47efbdfd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.926906] env[63372]: DEBUG oslo_vmware.api [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024920, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.932069] env[63372]: DEBUG oslo_vmware.api [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1159.932069] env[63372]: value = "task-1024921" [ 1159.932069] env[63372]: _type = "Task" [ 1159.932069] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.939162] env[63372]: DEBUG oslo_vmware.api [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024921, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.313247] env[63372]: DEBUG oslo_concurrency.lockutils [None req-704d420c-b279-441f-a95c-1a1d114d8cfc tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.313444] env[63372]: DEBUG oslo_concurrency.lockutils [None req-704d420c-b279-441f-a95c-1a1d114d8cfc tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquired lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.313623] env[63372]: DEBUG nova.network.neutron [None req-704d420c-b279-441f-a95c-1a1d114d8cfc tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1160.423970] env[63372]: DEBUG oslo_vmware.api [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024920, 'name': PowerOffVM_Task, 'duration_secs': 0.168261} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.424277] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1160.424454] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1160.424708] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1843b33-711b-498f-b619-b599918d54bb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.441058] env[63372]: DEBUG oslo_vmware.api [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024921, 'name': PowerOffVM_Task, 'duration_secs': 0.198684} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.441295] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1160.441463] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1160.441714] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6c14d617-5a15-4382-a30a-72e90059d464 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.504287] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1160.504541] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1160.504704] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Deleting the datastore file [datastore2] 1724ce03-c8e2-415d-a380-59ac69fbfb57 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1160.504971] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-733896bd-bc4a-49de-bfc2-94171a77eabc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.513228] env[63372]: DEBUG oslo_vmware.api [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1160.513228] env[63372]: value = "task-1024925" [ 1160.513228] env[63372]: _type = "Task" [ 1160.513228] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.517008] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1160.517222] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1160.517400] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleting the datastore file [datastore2] 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1160.517984] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-07e02279-d0e0-406a-ba8a-739615c2df31 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.522355] env[63372]: DEBUG oslo_vmware.api [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024925, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.526538] env[63372]: DEBUG oslo_vmware.api [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for the task: (returnval){ [ 1160.526538] env[63372]: value = "task-1024926" [ 1160.526538] env[63372]: _type = "Task" [ 1160.526538] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.533210] env[63372]: DEBUG oslo_vmware.api [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024926, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.023659] env[63372]: DEBUG oslo_vmware.api [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024925, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179695} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.024047] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1161.024130] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1161.024311] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1161.024493] env[63372]: INFO nova.compute.manager [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1161.024729] env[63372]: DEBUG oslo.service.loopingcall [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1161.024924] env[63372]: DEBUG nova.compute.manager [-] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1161.025025] env[63372]: DEBUG nova.network.neutron [-] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1161.027275] env[63372]: DEBUG nova.network.neutron [None req-704d420c-b279-441f-a95c-1a1d114d8cfc tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Updating instance_info_cache with network_info: [{"id": "146031c4-2f32-4085-9fc6-3060e7db9d08", "address": "fa:16:3e:70:3b:d5", "network": {"id": "01a8f000-169c-4202-a00a-04065baa3971", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-62305956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e46b6acae4a4b71bc390bd513ddd2f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d5d0e0d-cdec-474a-a891-a9ceff15a8b2", "external-id": "nsx-vlan-transportzone-456", "segmentation_id": 456, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap146031c4-2f", "ovs_interfaceid": "146031c4-2f32-4085-9fc6-3060e7db9d08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.036583] env[63372]: DEBUG oslo_vmware.api [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Task: {'id': task-1024926, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132956} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.036805] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1161.036988] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1161.037169] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1161.037339] env[63372]: INFO nova.compute.manager [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1161.037572] env[63372]: DEBUG oslo.service.loopingcall [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1161.038336] env[63372]: DEBUG nova.compute.manager [-] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1161.038422] env[63372]: DEBUG nova.network.neutron [-] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1161.510136] env[63372]: DEBUG nova.compute.manager [req-f328235c-d233-4c2b-ad1d-af9521d7f301 req-1ba16104-3933-4a91-8a22-1dd8ea4fd1ec service nova] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Received event network-vif-deleted-99872bc9-8786-4e6d-9978-42f880ba8407 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1161.510136] env[63372]: INFO nova.compute.manager [req-f328235c-d233-4c2b-ad1d-af9521d7f301 req-1ba16104-3933-4a91-8a22-1dd8ea4fd1ec service nova] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Neutron deleted interface 99872bc9-8786-4e6d-9978-42f880ba8407; detaching it from the instance and deleting it from the info cache [ 1161.510136] env[63372]: DEBUG nova.network.neutron [req-f328235c-d233-4c2b-ad1d-af9521d7f301 req-1ba16104-3933-4a91-8a22-1dd8ea4fd1ec service nova] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.523859] env[63372]: DEBUG nova.compute.manager [req-a08cb936-dd6a-4076-8ce2-55033f6a00e5 req-be2fb169-8ed0-482d-9543-5b36adcb3d74 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Received event network-vif-deleted-099091f6-c23b-48d0-9c21-76b4bcc3ad06 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1161.524071] env[63372]: INFO nova.compute.manager [req-a08cb936-dd6a-4076-8ce2-55033f6a00e5 req-be2fb169-8ed0-482d-9543-5b36adcb3d74 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Neutron deleted interface 099091f6-c23b-48d0-9c21-76b4bcc3ad06; detaching it from the instance and deleting it from the info cache [ 1161.524284] env[63372]: DEBUG nova.network.neutron [req-a08cb936-dd6a-4076-8ce2-55033f6a00e5 req-be2fb169-8ed0-482d-9543-5b36adcb3d74 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.532620] env[63372]: DEBUG oslo_concurrency.lockutils [None req-704d420c-b279-441f-a95c-1a1d114d8cfc tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Releasing lock "refresh_cache-802deb8b-978b-4d01-b19c-a0090e0029f0" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.534303] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b19478-b9aa-4999-9069-173eb6872fef {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.541307] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-704d420c-b279-441f-a95c-1a1d114d8cfc tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Resuming the VM {{(pid=63372) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1161.541794] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fff55505-c8e8-4099-9e9e-be719bf47c43 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.547858] env[63372]: DEBUG oslo_vmware.api [None req-704d420c-b279-441f-a95c-1a1d114d8cfc tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1161.547858] env[63372]: value = "task-1024927" [ 1161.547858] env[63372]: _type = "Task" [ 1161.547858] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.555233] env[63372]: DEBUG oslo_vmware.api [None req-704d420c-b279-441f-a95c-1a1d114d8cfc tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024927, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.981356] env[63372]: DEBUG nova.network.neutron [-] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.003037] env[63372]: DEBUG nova.network.neutron [-] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.013423] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4254faaf-e888-4669-ac5f-245da160648f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.025275] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba86a941-4f9e-4c48-9e69-77c18dc0f664 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.036655] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-03df3d88-a626-4460-aa25-60836b18b5e3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.046687] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1962d8da-ec11-4953-bdf5-0fa7c76ed5e3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.064118] env[63372]: DEBUG nova.compute.manager [req-f328235c-d233-4c2b-ad1d-af9521d7f301 req-1ba16104-3933-4a91-8a22-1dd8ea4fd1ec service nova] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Detach interface failed, port_id=99872bc9-8786-4e6d-9978-42f880ba8407, reason: Instance 1724ce03-c8e2-415d-a380-59ac69fbfb57 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1162.072625] env[63372]: DEBUG oslo_vmware.api [None req-704d420c-b279-441f-a95c-1a1d114d8cfc tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024927, 'name': PowerOnVM_Task, 'duration_secs': 0.450928} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.072934] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-704d420c-b279-441f-a95c-1a1d114d8cfc tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Resumed the VM {{(pid=63372) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1162.073151] env[63372]: DEBUG nova.compute.manager [None req-704d420c-b279-441f-a95c-1a1d114d8cfc tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1162.073930] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e146d9b6-96fd-4d0e-b198-073b8f9167f0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.083259] env[63372]: DEBUG nova.compute.manager [req-a08cb936-dd6a-4076-8ce2-55033f6a00e5 req-be2fb169-8ed0-482d-9543-5b36adcb3d74 service nova] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Detach interface failed, port_id=099091f6-c23b-48d0-9c21-76b4bcc3ad06, reason: Instance 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1162.485623] env[63372]: INFO nova.compute.manager [-] [instance: 1724ce03-c8e2-415d-a380-59ac69fbfb57] Took 1.46 seconds to deallocate network for instance. [ 1162.505434] env[63372]: INFO nova.compute.manager [-] [instance: 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b] Took 1.47 seconds to deallocate network for instance. [ 1162.532417] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Volume attach. Driver type: vmdk {{(pid=63372) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1162.532700] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227479', 'volume_id': '598ca7c3-f9fb-4fe4-939a-9ca37fc19323', 'name': 'volume-598ca7c3-f9fb-4fe4-939a-9ca37fc19323', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'fafed45e-c9dc-4196-883f-a23570fd0583', 'attached_at': '', 'detached_at': '', 'volume_id': '598ca7c3-f9fb-4fe4-939a-9ca37fc19323', 'serial': '598ca7c3-f9fb-4fe4-939a-9ca37fc19323'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1162.533833] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b760c2c-8440-4d00-979e-536aa330fd8b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.550991] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab2afbc-56ff-42a4-9936-0f239d0d1df4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.574971] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] volume-598ca7c3-f9fb-4fe4-939a-9ca37fc19323/volume-598ca7c3-f9fb-4fe4-939a-9ca37fc19323.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1162.575295] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e002d084-f0dd-4281-9ab4-8adb8170b979 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.596108] env[63372]: DEBUG oslo_vmware.api [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1162.596108] env[63372]: value = "task-1024928" [ 1162.596108] env[63372]: _type = "Task" [ 1162.596108] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.606751] env[63372]: DEBUG oslo_vmware.api [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024928, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.991817] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.992150] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.992381] env[63372]: DEBUG nova.objects.instance [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lazy-loading 'resources' on Instance uuid 1724ce03-c8e2-415d-a380-59ac69fbfb57 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1163.011083] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.107060] env[63372]: DEBUG oslo_vmware.api [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024928, 'name': ReconfigVM_Task, 'duration_secs': 0.352322} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.107515] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Reconfigured VM instance instance-0000006f to attach disk [datastore1] volume-598ca7c3-f9fb-4fe4-939a-9ca37fc19323/volume-598ca7c3-f9fb-4fe4-939a-9ca37fc19323.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1163.111988] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6955d75d-38c5-4706-bf58-2ae37551bd4c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.126480] env[63372]: DEBUG oslo_vmware.api [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1163.126480] env[63372]: value = "task-1024929" [ 1163.126480] env[63372]: _type = "Task" [ 1163.126480] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.135011] env[63372]: DEBUG oslo_vmware.api [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024929, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.180014] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "802deb8b-978b-4d01-b19c-a0090e0029f0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.180327] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "802deb8b-978b-4d01-b19c-a0090e0029f0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.180549] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "802deb8b-978b-4d01-b19c-a0090e0029f0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.180737] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "802deb8b-978b-4d01-b19c-a0090e0029f0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.180914] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "802deb8b-978b-4d01-b19c-a0090e0029f0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.183489] env[63372]: INFO nova.compute.manager [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Terminating instance [ 1163.185321] env[63372]: DEBUG nova.compute.manager [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1163.185522] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1163.186344] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda9fa66-1c31-4ffc-9025-622e9dcf6819 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.194012] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1163.194277] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afcafbd9-b700-4a7e-a067-df927dadc24d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.200969] env[63372]: DEBUG oslo_vmware.api [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1163.200969] env[63372]: value = "task-1024930" [ 1163.200969] env[63372]: _type = "Task" [ 1163.200969] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.208416] env[63372]: DEBUG oslo_vmware.api [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024930, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.578397] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e31205-ded1-44e6-8566-f32b5fdc23af {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.585749] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb31aca-e6a4-4531-bd42-eb6809868705 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.614801] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e743335b-5f75-41c8-a7dd-5398772edad8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.621435] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdcc73fb-d2a8-4a00-afe9-e951dedc4961 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.637488] env[63372]: DEBUG nova.compute.provider_tree [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1163.643163] env[63372]: DEBUG oslo_vmware.api [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024929, 'name': ReconfigVM_Task, 'duration_secs': 0.201016} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.643313] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227479', 'volume_id': '598ca7c3-f9fb-4fe4-939a-9ca37fc19323', 'name': 'volume-598ca7c3-f9fb-4fe4-939a-9ca37fc19323', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'fafed45e-c9dc-4196-883f-a23570fd0583', 'attached_at': '', 'detached_at': '', 'volume_id': '598ca7c3-f9fb-4fe4-939a-9ca37fc19323', 'serial': '598ca7c3-f9fb-4fe4-939a-9ca37fc19323'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1163.710430] env[63372]: DEBUG oslo_vmware.api [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024930, 'name': PowerOffVM_Task, 'duration_secs': 0.204645} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.710684] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1163.710852] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1163.711116] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4bd6649b-3a15-453b-9424-88a90808ef3a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.780185] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1163.780613] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1163.780852] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleting the datastore file [datastore2] 802deb8b-978b-4d01-b19c-a0090e0029f0 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1163.781127] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-781d2fa0-3104-4f51-bdc9-b414496e570f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.788174] env[63372]: DEBUG oslo_vmware.api [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for the task: (returnval){ [ 1163.788174] env[63372]: value = "task-1024932" [ 1163.788174] env[63372]: _type = "Task" [ 1163.788174] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.795489] env[63372]: DEBUG oslo_vmware.api [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024932, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.140960] env[63372]: DEBUG nova.scheduler.client.report [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1164.298709] env[63372]: DEBUG oslo_vmware.api [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Task: {'id': task-1024932, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145784} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.300049] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.300049] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1164.300049] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1164.300049] env[63372]: INFO nova.compute.manager [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1164.300049] env[63372]: DEBUG oslo.service.loopingcall [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1164.300297] env[63372]: DEBUG nova.compute.manager [-] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1164.300338] env[63372]: DEBUG nova.network.neutron [-] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1164.645794] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.654s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.649292] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.637s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.649572] env[63372]: DEBUG nova.objects.instance [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lazy-loading 'resources' on Instance uuid 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.673779] env[63372]: INFO nova.scheduler.client.report [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Deleted allocations for instance 1724ce03-c8e2-415d-a380-59ac69fbfb57 [ 1164.678011] env[63372]: DEBUG nova.objects.instance [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lazy-loading 'flavor' on Instance uuid fafed45e-c9dc-4196-883f-a23570fd0583 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.781664] env[63372]: DEBUG nova.compute.manager [req-53a2a92c-5663-4d87-8b2e-29fe10dd8bcc req-ba3f8c58-1e57-4704-87b8-161a6eb19d8f service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Received event network-vif-deleted-146031c4-2f32-4085-9fc6-3060e7db9d08 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1164.781916] env[63372]: INFO nova.compute.manager [req-53a2a92c-5663-4d87-8b2e-29fe10dd8bcc req-ba3f8c58-1e57-4704-87b8-161a6eb19d8f service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Neutron deleted interface 146031c4-2f32-4085-9fc6-3060e7db9d08; detaching it from the instance and deleting it from the info cache [ 1164.782316] env[63372]: DEBUG nova.network.neutron [req-53a2a92c-5663-4d87-8b2e-29fe10dd8bcc req-ba3f8c58-1e57-4704-87b8-161a6eb19d8f service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.183615] env[63372]: DEBUG oslo_concurrency.lockutils [None req-bf876115-3539-47ee-9c7c-a718e6b6107b tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "1724ce03-c8e2-415d-a380-59ac69fbfb57" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.279s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.184642] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9581fe85-dd3c-4836-837a-51b486c614e5 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.751s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.246629] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591d5e67-29db-4783-91a7-fb38bfc595b8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.255451] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e28bf8c-4a32-421c-970d-9efe265b5c6a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.258806] env[63372]: DEBUG nova.network.neutron [-] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.288835] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-60e0df65-1228-45dc-802f-eca49bcf2c82 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.292308] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74021c7d-860c-4c62-a46b-43fe40f0cce1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.303434] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c32de17f-7a2a-4915-a529-e33070e48c00 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.311873] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de240b48-2677-4a7a-8694-3385a34d5500 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.334350] env[63372]: DEBUG nova.compute.provider_tree [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1165.344311] env[63372]: DEBUG nova.compute.manager [req-53a2a92c-5663-4d87-8b2e-29fe10dd8bcc req-ba3f8c58-1e57-4704-87b8-161a6eb19d8f service nova] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Detach interface failed, port_id=146031c4-2f32-4085-9fc6-3060e7db9d08, reason: Instance 802deb8b-978b-4d01-b19c-a0090e0029f0 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1165.348104] env[63372]: DEBUG nova.scheduler.client.report [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1165.525070] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.525375] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.525599] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.525786] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.525964] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.528315] env[63372]: INFO nova.compute.manager [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Terminating instance [ 1165.530221] env[63372]: DEBUG nova.compute.manager [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1165.530432] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1165.531308] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5269738-4354-4809-a742-5c68bc2087a6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.539316] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1165.539566] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-90c9759c-8771-4b5f-9ea3-6192f7ba45d7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.547932] env[63372]: DEBUG oslo_vmware.api [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1165.547932] env[63372]: value = "task-1024933" [ 1165.547932] env[63372]: _type = "Task" [ 1165.547932] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.556508] env[63372]: DEBUG oslo_vmware.api [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024933, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.761779] env[63372]: INFO nova.compute.manager [-] [instance: 802deb8b-978b-4d01-b19c-a0090e0029f0] Took 1.46 seconds to deallocate network for instance. [ 1165.825690] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6fb104d4-faa2-4102-aa4a-3b261370df25 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "fafed45e-c9dc-4196-883f-a23570fd0583" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.825975] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6fb104d4-faa2-4102-aa4a-3b261370df25 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.826198] env[63372]: DEBUG nova.compute.manager [None req-6fb104d4-faa2-4102-aa4a-3b261370df25 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1165.827080] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb6b155-5abe-4f74-a1bf-d93214bde5fd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.834465] env[63372]: DEBUG nova.compute.manager [None req-6fb104d4-faa2-4102-aa4a-3b261370df25 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63372) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1165.835070] env[63372]: DEBUG nova.objects.instance [None req-6fb104d4-faa2-4102-aa4a-3b261370df25 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lazy-loading 'flavor' on Instance uuid fafed45e-c9dc-4196-883f-a23570fd0583 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.849647] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.201s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.868684] env[63372]: INFO nova.scheduler.client.report [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Deleted allocations for instance 2b907d6c-ee2c-4cd5-a77a-babfeb41b88b [ 1166.058470] env[63372]: DEBUG oslo_vmware.api [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024933, 'name': PowerOffVM_Task, 'duration_secs': 0.196753} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.058803] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1166.059024] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1166.059303] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6bc8afdd-b143-4f7a-95fa-7b70dbdedc2a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.125274] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1166.125568] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1166.125694] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Deleting the datastore file [datastore2] ddca7aa5-fcc8-4835-bfdb-47781335f2cd {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1166.125960] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6fc0e692-e40a-46b5-9253-b10893a1df44 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.132745] env[63372]: DEBUG oslo_vmware.api [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for the task: (returnval){ [ 1166.132745] env[63372]: value = "task-1024936" [ 1166.132745] env[63372]: _type = "Task" [ 1166.132745] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.141249] env[63372]: DEBUG oslo_vmware.api [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024936, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.268334] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.268637] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.268862] env[63372]: DEBUG nova.objects.instance [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lazy-loading 'resources' on Instance uuid 802deb8b-978b-4d01-b19c-a0090e0029f0 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.340263] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb104d4-faa2-4102-aa4a-3b261370df25 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1166.340523] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-450f14d7-3d31-4dd5-8435-bd26f5ad9a8a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.348391] env[63372]: DEBUG oslo_vmware.api [None req-6fb104d4-faa2-4102-aa4a-3b261370df25 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1166.348391] env[63372]: value = "task-1024937" [ 1166.348391] env[63372]: _type = "Task" [ 1166.348391] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.356833] env[63372]: DEBUG oslo_vmware.api [None req-6fb104d4-faa2-4102-aa4a-3b261370df25 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024937, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.375359] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9fbb4167-4c2b-4944-857d-5bf2f0a63a23 tempest-ServerActionsTestOtherB-1719885972 tempest-ServerActionsTestOtherB-1719885972-project-member] Lock "2b907d6c-ee2c-4cd5-a77a-babfeb41b88b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.485s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.642646] env[63372]: DEBUG oslo_vmware.api [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Task: {'id': task-1024936, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138792} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.642931] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1166.643141] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1166.643392] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1166.643589] env[63372]: INFO nova.compute.manager [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1166.643836] env[63372]: DEBUG oslo.service.loopingcall [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1166.644048] env[63372]: DEBUG nova.compute.manager [-] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1166.644145] env[63372]: DEBUG nova.network.neutron [-] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1166.855226] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69573d16-147c-4340-8c54-6fe8ec2beb57 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.860356] env[63372]: DEBUG oslo_vmware.api [None req-6fb104d4-faa2-4102-aa4a-3b261370df25 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024937, 'name': PowerOffVM_Task, 'duration_secs': 0.1936} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.861028] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb104d4-faa2-4102-aa4a-3b261370df25 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1166.861193] env[63372]: DEBUG nova.compute.manager [None req-6fb104d4-faa2-4102-aa4a-3b261370df25 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1166.862209] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69ca06a0-c020-463a-b893-e875af11a9b9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.868678] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3515c57-b13a-4f9c-8c7f-f5fe8045b1b8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.904928] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-901a5bf7-879f-4ff1-8b99-952f9b2df78c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.908369] env[63372]: DEBUG nova.compute.manager [req-6c881a1f-18a2-422c-831b-9562b6b7349e req-7ba21bd4-e0af-49c0-bf49-be13ed8a0671 service nova] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Received event network-vif-deleted-b54d2aaa-a37b-4db7-b735-f9d2a6a9534b {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1166.908560] env[63372]: INFO nova.compute.manager [req-6c881a1f-18a2-422c-831b-9562b6b7349e req-7ba21bd4-e0af-49c0-bf49-be13ed8a0671 service nova] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Neutron deleted interface b54d2aaa-a37b-4db7-b735-f9d2a6a9534b; detaching it from the instance and deleting it from the info cache [ 1166.908730] env[63372]: DEBUG nova.network.neutron [req-6c881a1f-18a2-422c-831b-9562b6b7349e req-7ba21bd4-e0af-49c0-bf49-be13ed8a0671 service nova] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.916353] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b95b404f-e6e4-40cb-9d92-02d42bda2d87 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.930864] env[63372]: DEBUG nova.compute.provider_tree [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.378865] env[63372]: DEBUG oslo_concurrency.lockutils [None req-6fb104d4-faa2-4102-aa4a-3b261370df25 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.553s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.379750] env[63372]: DEBUG nova.network.neutron [-] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.412175] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-95f6cd7b-2fda-406a-a542-3f589746bc54 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.423575] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-644c1190-5d96-4ad3-a435-532a41720833 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.434793] env[63372]: DEBUG nova.scheduler.client.report [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1167.455380] env[63372]: DEBUG nova.compute.manager [req-6c881a1f-18a2-422c-831b-9562b6b7349e req-7ba21bd4-e0af-49c0-bf49-be13ed8a0671 service nova] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Detach interface failed, port_id=b54d2aaa-a37b-4db7-b735-f9d2a6a9534b, reason: Instance ddca7aa5-fcc8-4835-bfdb-47781335f2cd could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1167.884794] env[63372]: INFO nova.compute.manager [-] [instance: ddca7aa5-fcc8-4835-bfdb-47781335f2cd] Took 1.24 seconds to deallocate network for instance. [ 1167.939289] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.670s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.959151] env[63372]: INFO nova.scheduler.client.report [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Deleted allocations for instance 802deb8b-978b-4d01-b19c-a0090e0029f0 [ 1168.229786] env[63372]: DEBUG nova.objects.instance [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lazy-loading 'flavor' on Instance uuid fafed45e-c9dc-4196-883f-a23570fd0583 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.390969] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.391273] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.391515] env[63372]: DEBUG nova.objects.instance [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lazy-loading 'resources' on Instance uuid ddca7aa5-fcc8-4835-bfdb-47781335f2cd {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.466448] env[63372]: DEBUG oslo_concurrency.lockutils [None req-7d66ac5a-2735-4160-92c1-80a591720dea tempest-ServerActionsTestJSON-385763715 tempest-ServerActionsTestJSON-385763715-project-member] Lock "802deb8b-978b-4d01-b19c-a0090e0029f0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.286s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.734324] env[63372]: DEBUG oslo_concurrency.lockutils [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1168.734556] env[63372]: DEBUG oslo_concurrency.lockutils [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquired lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.734654] env[63372]: DEBUG nova.network.neutron [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1168.734834] env[63372]: DEBUG nova.objects.instance [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lazy-loading 'info_cache' on Instance uuid fafed45e-c9dc-4196-883f-a23570fd0583 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.962046] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ddc40f7-f2fa-418a-97ee-d41f046fbeb4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.968709] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4040975-6ad7-48fa-87dd-af14c7ffa2bc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.001647] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15724a38-663e-4227-8176-59f2297d9be3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.013265] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6d407a-8a12-4de8-a745-b4d75a727db8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.026351] env[63372]: DEBUG nova.compute.provider_tree [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1169.238465] env[63372]: DEBUG nova.objects.base [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1169.529320] env[63372]: DEBUG nova.scheduler.client.report [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1170.036245] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.645s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.067166] env[63372]: INFO nova.scheduler.client.report [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Deleted allocations for instance ddca7aa5-fcc8-4835-bfdb-47781335f2cd [ 1170.121965] env[63372]: DEBUG nova.network.neutron [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Updating instance_info_cache with network_info: [{"id": "7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db", "address": "fa:16:3e:94:43:d7", "network": {"id": "b161f533-292b-431a-a171-e9d61bc90105", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1928954423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4b088060a2c4159a62e40e7e627076a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ca80379-c6", "ovs_interfaceid": "7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1170.577594] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebe9bed7-4f41-4e5c-9b5b-d844a131cd4e tempest-ServerRescueNegativeTestJSON-922384901 tempest-ServerRescueNegativeTestJSON-922384901-project-member] Lock "ddca7aa5-fcc8-4835-bfdb-47781335f2cd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.050s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.626143] env[63372]: DEBUG oslo_concurrency.lockutils [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Releasing lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1171.131662] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1171.132042] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c702cb09-a0f5-4056-bd2f-9adf707d001f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.143036] env[63372]: DEBUG oslo_vmware.api [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1171.143036] env[63372]: value = "task-1024938" [ 1171.143036] env[63372]: _type = "Task" [ 1171.143036] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.151546] env[63372]: DEBUG oslo_vmware.api [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024938, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.652060] env[63372]: DEBUG oslo_vmware.api [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024938, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.151949] env[63372]: DEBUG oslo_vmware.api [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024938, 'name': PowerOnVM_Task, 'duration_secs': 0.596452} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.152296] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1172.152500] env[63372]: DEBUG nova.compute.manager [None req-afbb7f2b-4f8e-41d7-90b0-26f61e6e3e7f tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1172.153319] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deb9f6ec-4483-418f-9f4a-c0156c55acc4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.911739] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Acquiring lock "ea851e4f-0ccc-487c-ac4a-2d658a9337ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.913562] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Lock "ea851e4f-0ccc-487c-ac4a-2d658a9337ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.415832] env[63372]: DEBUG nova.compute.manager [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1175.939634] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.939931] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.941494] env[63372]: INFO nova.compute.claims [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1177.007292] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1254a0b-0e95-4cf9-86fa-c7be8cfc0327 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.014830] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef5fe0e-deaa-4f4a-9e56-1417d9b1a3f0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.045050] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d923c6a6-1117-4dcd-b4e2-d5d78aca0708 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.051844] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897725ad-d650-45ca-be24-3407153a90ee {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.065611] env[63372]: DEBUG nova.compute.provider_tree [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1177.568655] env[63372]: DEBUG nova.scheduler.client.report [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1178.073421] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.133s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.073967] env[63372]: DEBUG nova.compute.manager [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1178.579395] env[63372]: DEBUG nova.compute.utils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1178.581100] env[63372]: DEBUG nova.compute.manager [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1178.581330] env[63372]: DEBUG nova.network.neutron [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1178.630335] env[63372]: DEBUG nova.policy [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b1b96c47f324467ac9ba883a7a6c3f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '797d74504a6e4e4f8aedd36c05dcd6a7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1178.861144] env[63372]: DEBUG nova.network.neutron [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Successfully created port: 60967e30-46f4-4313-b1dc-5df97f4eef71 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1179.084878] env[63372]: DEBUG nova.compute.manager [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1180.096344] env[63372]: DEBUG nova.compute.manager [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1180.123622] env[63372]: DEBUG nova.virt.hardware [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1180.123891] env[63372]: DEBUG nova.virt.hardware [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1180.124085] env[63372]: DEBUG nova.virt.hardware [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1180.124295] env[63372]: DEBUG nova.virt.hardware [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1180.124460] env[63372]: DEBUG nova.virt.hardware [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1180.124618] env[63372]: DEBUG nova.virt.hardware [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1180.124829] env[63372]: DEBUG nova.virt.hardware [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1180.125029] env[63372]: DEBUG nova.virt.hardware [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1180.125244] env[63372]: DEBUG nova.virt.hardware [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1180.125421] env[63372]: DEBUG nova.virt.hardware [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1180.125599] env[63372]: DEBUG nova.virt.hardware [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1180.126535] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0499dd-b41c-4c00-908c-89499f4d1552 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.134841] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367662ee-4bbb-413b-b99d-72a2ff38d160 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.240171] env[63372]: DEBUG nova.compute.manager [req-e1d8731a-7d76-484f-89d0-e81a872fb6ac req-41f4a092-6251-4590-9ba0-be3c28f084ea service nova] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Received event network-vif-plugged-60967e30-46f4-4313-b1dc-5df97f4eef71 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1180.240171] env[63372]: DEBUG oslo_concurrency.lockutils [req-e1d8731a-7d76-484f-89d0-e81a872fb6ac req-41f4a092-6251-4590-9ba0-be3c28f084ea service nova] Acquiring lock "ea851e4f-0ccc-487c-ac4a-2d658a9337ae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1180.240171] env[63372]: DEBUG oslo_concurrency.lockutils [req-e1d8731a-7d76-484f-89d0-e81a872fb6ac req-41f4a092-6251-4590-9ba0-be3c28f084ea service nova] Lock "ea851e4f-0ccc-487c-ac4a-2d658a9337ae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1180.240171] env[63372]: DEBUG oslo_concurrency.lockutils [req-e1d8731a-7d76-484f-89d0-e81a872fb6ac req-41f4a092-6251-4590-9ba0-be3c28f084ea service nova] Lock "ea851e4f-0ccc-487c-ac4a-2d658a9337ae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.240171] env[63372]: DEBUG nova.compute.manager [req-e1d8731a-7d76-484f-89d0-e81a872fb6ac req-41f4a092-6251-4590-9ba0-be3c28f084ea service nova] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] No waiting events found dispatching network-vif-plugged-60967e30-46f4-4313-b1dc-5df97f4eef71 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1180.240171] env[63372]: WARNING nova.compute.manager [req-e1d8731a-7d76-484f-89d0-e81a872fb6ac req-41f4a092-6251-4590-9ba0-be3c28f084ea service nova] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Received unexpected event network-vif-plugged-60967e30-46f4-4313-b1dc-5df97f4eef71 for instance with vm_state building and task_state spawning. [ 1180.317816] env[63372]: DEBUG nova.network.neutron [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Successfully updated port: 60967e30-46f4-4313-b1dc-5df97f4eef71 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1180.819920] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Acquiring lock "refresh_cache-ea851e4f-0ccc-487c-ac4a-2d658a9337ae" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1180.820122] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Acquired lock "refresh_cache-ea851e4f-0ccc-487c-ac4a-2d658a9337ae" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1180.820282] env[63372]: DEBUG nova.network.neutron [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1181.350935] env[63372]: DEBUG nova.network.neutron [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1181.480778] env[63372]: DEBUG nova.network.neutron [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Updating instance_info_cache with network_info: [{"id": "60967e30-46f4-4313-b1dc-5df97f4eef71", "address": "fa:16:3e:21:01:22", "network": {"id": "6dc2e2e1-de52-4950-bece-da80183a3d7b", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1871295198-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "797d74504a6e4e4f8aedd36c05dcd6a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "438671d0-9468-4e44-84c1-4c0ebaa743e0", "external-id": "nsx-vlan-transportzone-918", "segmentation_id": 918, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60967e30-46", "ovs_interfaceid": "60967e30-46f4-4313-b1dc-5df97f4eef71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.983374] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Releasing lock "refresh_cache-ea851e4f-0ccc-487c-ac4a-2d658a9337ae" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.983690] env[63372]: DEBUG nova.compute.manager [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Instance network_info: |[{"id": "60967e30-46f4-4313-b1dc-5df97f4eef71", "address": "fa:16:3e:21:01:22", "network": {"id": "6dc2e2e1-de52-4950-bece-da80183a3d7b", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1871295198-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "797d74504a6e4e4f8aedd36c05dcd6a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "438671d0-9468-4e44-84c1-4c0ebaa743e0", "external-id": "nsx-vlan-transportzone-918", "segmentation_id": 918, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60967e30-46", "ovs_interfaceid": "60967e30-46f4-4313-b1dc-5df97f4eef71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1181.984154] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:01:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '438671d0-9468-4e44-84c1-4c0ebaa743e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '60967e30-46f4-4313-b1dc-5df97f4eef71', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1181.992050] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Creating folder: Project (797d74504a6e4e4f8aedd36c05dcd6a7). Parent ref: group-v227230. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1181.992050] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d3468607-0e23-4489-9ceb-354bcc13889a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.004658] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Created folder: Project (797d74504a6e4e4f8aedd36c05dcd6a7) in parent group-v227230. [ 1182.004844] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Creating folder: Instances. Parent ref: group-v227480. {{(pid=63372) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1182.005135] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3bfac590-9899-449d-a490-61c505749c83 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.014335] env[63372]: INFO nova.virt.vmwareapi.vm_util [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Created folder: Instances in parent group-v227480. [ 1182.014568] env[63372]: DEBUG oslo.service.loopingcall [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1182.014751] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1182.014942] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd6c0a0e-0ac2-43c8-ae06-dc1aaed45bb5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.033316] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1182.033316] env[63372]: value = "task-1024941" [ 1182.033316] env[63372]: _type = "Task" [ 1182.033316] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.040300] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024941, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.265599] env[63372]: DEBUG nova.compute.manager [req-6f904124-221c-46b8-a51f-f3128a9f756d req-e7230d81-5ecd-4f41-9a80-ebd1170bbb9c service nova] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Received event network-changed-60967e30-46f4-4313-b1dc-5df97f4eef71 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1182.265808] env[63372]: DEBUG nova.compute.manager [req-6f904124-221c-46b8-a51f-f3128a9f756d req-e7230d81-5ecd-4f41-9a80-ebd1170bbb9c service nova] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Refreshing instance network info cache due to event network-changed-60967e30-46f4-4313-b1dc-5df97f4eef71. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1182.266015] env[63372]: DEBUG oslo_concurrency.lockutils [req-6f904124-221c-46b8-a51f-f3128a9f756d req-e7230d81-5ecd-4f41-9a80-ebd1170bbb9c service nova] Acquiring lock "refresh_cache-ea851e4f-0ccc-487c-ac4a-2d658a9337ae" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.266175] env[63372]: DEBUG oslo_concurrency.lockutils [req-6f904124-221c-46b8-a51f-f3128a9f756d req-e7230d81-5ecd-4f41-9a80-ebd1170bbb9c service nova] Acquired lock "refresh_cache-ea851e4f-0ccc-487c-ac4a-2d658a9337ae" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.266338] env[63372]: DEBUG nova.network.neutron [req-6f904124-221c-46b8-a51f-f3128a9f756d req-e7230d81-5ecd-4f41-9a80-ebd1170bbb9c service nova] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Refreshing network info cache for port 60967e30-46f4-4313-b1dc-5df97f4eef71 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1182.543685] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024941, 'name': CreateVM_Task, 'duration_secs': 0.31909} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.544101] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1182.544515] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.544680] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.544995] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1182.545248] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67ad4fed-52aa-423b-acd7-10c2a64ac4d9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.549444] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Waiting for the task: (returnval){ [ 1182.549444] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]529e3c87-1898-a9cf-ad1d-4dacf6de9c24" [ 1182.549444] env[63372]: _type = "Task" [ 1182.549444] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.556856] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529e3c87-1898-a9cf-ad1d-4dacf6de9c24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.949548] env[63372]: DEBUG nova.network.neutron [req-6f904124-221c-46b8-a51f-f3128a9f756d req-e7230d81-5ecd-4f41-9a80-ebd1170bbb9c service nova] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Updated VIF entry in instance network info cache for port 60967e30-46f4-4313-b1dc-5df97f4eef71. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1182.949936] env[63372]: DEBUG nova.network.neutron [req-6f904124-221c-46b8-a51f-f3128a9f756d req-e7230d81-5ecd-4f41-9a80-ebd1170bbb9c service nova] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Updating instance_info_cache with network_info: [{"id": "60967e30-46f4-4313-b1dc-5df97f4eef71", "address": "fa:16:3e:21:01:22", "network": {"id": "6dc2e2e1-de52-4950-bece-da80183a3d7b", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1871295198-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "797d74504a6e4e4f8aedd36c05dcd6a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "438671d0-9468-4e44-84c1-4c0ebaa743e0", "external-id": "nsx-vlan-transportzone-918", "segmentation_id": 918, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60967e30-46", "ovs_interfaceid": "60967e30-46f4-4313-b1dc-5df97f4eef71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1183.059690] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]529e3c87-1898-a9cf-ad1d-4dacf6de9c24, 'name': SearchDatastore_Task, 'duration_secs': 0.011085} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.059934] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.060185] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1183.060420] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.060572] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.060752] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1183.061009] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b5a8dd52-e064-487a-af13-21ba58c691ec {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.068548] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1183.068720] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1183.069393] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b32c28a1-88a8-4b76-9d72-9770fac9e0ca {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.073910] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Waiting for the task: (returnval){ [ 1183.073910] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]5266a67f-fcfc-c787-a82f-3ff0345873f1" [ 1183.073910] env[63372]: _type = "Task" [ 1183.073910] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.080890] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5266a67f-fcfc-c787-a82f-3ff0345873f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.452791] env[63372]: DEBUG oslo_concurrency.lockutils [req-6f904124-221c-46b8-a51f-f3128a9f756d req-e7230d81-5ecd-4f41-9a80-ebd1170bbb9c service nova] Releasing lock "refresh_cache-ea851e4f-0ccc-487c-ac4a-2d658a9337ae" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.584576] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]5266a67f-fcfc-c787-a82f-3ff0345873f1, 'name': SearchDatastore_Task, 'duration_secs': 0.007894} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.585381] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73753582-74ba-4619-b53a-5b1ba1c9c4ff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.590026] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Waiting for the task: (returnval){ [ 1183.590026] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52ea8f7b-9c4f-ccae-2129-91bf04fc0377" [ 1183.590026] env[63372]: _type = "Task" [ 1183.590026] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.596988] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ea8f7b-9c4f-ccae-2129-91bf04fc0377, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.100813] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52ea8f7b-9c4f-ccae-2129-91bf04fc0377, 'name': SearchDatastore_Task, 'duration_secs': 0.009335} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.101098] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.101363] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] ea851e4f-0ccc-487c-ac4a-2d658a9337ae/ea851e4f-0ccc-487c-ac4a-2d658a9337ae.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1184.101619] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c44ddecf-93c6-4e86-89e5-a2e4c0ae869f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.108959] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Waiting for the task: (returnval){ [ 1184.108959] env[63372]: value = "task-1024942" [ 1184.108959] env[63372]: _type = "Task" [ 1184.108959] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.116771] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': task-1024942, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.619235] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': task-1024942, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435996} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.619627] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] ea851e4f-0ccc-487c-ac4a-2d658a9337ae/ea851e4f-0ccc-487c-ac4a-2d658a9337ae.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1184.619786] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1184.619972] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9f39cab6-f771-484b-a4de-e499699c858a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.626093] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Waiting for the task: (returnval){ [ 1184.626093] env[63372]: value = "task-1024943" [ 1184.626093] env[63372]: _type = "Task" [ 1184.626093] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.632957] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': task-1024943, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.702677] env[63372]: DEBUG oslo_concurrency.lockutils [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "1e7590de-55ae-4725-be5a-e3435d98151a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.702965] env[63372]: DEBUG oslo_concurrency.lockutils [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "1e7590de-55ae-4725-be5a-e3435d98151a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.703230] env[63372]: DEBUG oslo_concurrency.lockutils [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "1e7590de-55ae-4725-be5a-e3435d98151a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.703431] env[63372]: DEBUG oslo_concurrency.lockutils [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "1e7590de-55ae-4725-be5a-e3435d98151a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.703604] env[63372]: DEBUG oslo_concurrency.lockutils [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "1e7590de-55ae-4725-be5a-e3435d98151a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.705868] env[63372]: INFO nova.compute.manager [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Terminating instance [ 1184.707692] env[63372]: DEBUG nova.compute.manager [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1184.707884] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1184.708702] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5662f128-5864-42fb-99a7-a07446f3906b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.716156] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1184.716378] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-91330361-e5a0-4e75-a8b9-8925852c2d75 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.723356] env[63372]: DEBUG oslo_vmware.api [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1184.723356] env[63372]: value = "task-1024944" [ 1184.723356] env[63372]: _type = "Task" [ 1184.723356] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.731403] env[63372]: DEBUG oslo_vmware.api [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024944, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.136329] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': task-1024943, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117834} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.136607] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1185.137377] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f2b486-59de-4ca8-a5f7-2fff26fca998 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.160375] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] ea851e4f-0ccc-487c-ac4a-2d658a9337ae/ea851e4f-0ccc-487c-ac4a-2d658a9337ae.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1185.160691] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ebe84154-f06a-4112-a886-3ff704207e53 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.180994] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Waiting for the task: (returnval){ [ 1185.180994] env[63372]: value = "task-1024945" [ 1185.180994] env[63372]: _type = "Task" [ 1185.180994] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.191668] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': task-1024945, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.232768] env[63372]: DEBUG oslo_vmware.api [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024944, 'name': PowerOffVM_Task, 'duration_secs': 0.196311} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.233029] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1185.233228] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1185.233482] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc7ffff8-89ae-4cf2-86dc-30b1f23c6397 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.308866] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1185.309105] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1185.309297] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Deleting the datastore file [datastore1] 1e7590de-55ae-4725-be5a-e3435d98151a {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1185.309574] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e6b8821-11f2-4dca-a185-f6053c1f2edf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.315763] env[63372]: DEBUG oslo_vmware.api [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1185.315763] env[63372]: value = "task-1024947" [ 1185.315763] env[63372]: _type = "Task" [ 1185.315763] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.323065] env[63372]: DEBUG oslo_vmware.api [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024947, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.691405] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': task-1024945, 'name': ReconfigVM_Task, 'duration_secs': 0.304773} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.691759] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Reconfigured VM instance instance-00000071 to attach disk [datastore2] ea851e4f-0ccc-487c-ac4a-2d658a9337ae/ea851e4f-0ccc-487c-ac4a-2d658a9337ae.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1185.692185] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fb2ca723-d5cd-4a0b-9170-56dad4e3c83a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.698946] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Waiting for the task: (returnval){ [ 1185.698946] env[63372]: value = "task-1024948" [ 1185.698946] env[63372]: _type = "Task" [ 1185.698946] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.706450] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': task-1024948, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.826096] env[63372]: DEBUG oslo_vmware.api [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024947, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145811} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.826403] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1185.826614] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1185.826797] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1185.827029] env[63372]: INFO nova.compute.manager [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1185.827305] env[63372]: DEBUG oslo.service.loopingcall [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1185.827504] env[63372]: DEBUG nova.compute.manager [-] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1185.827612] env[63372]: DEBUG nova.network.neutron [-] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1186.208900] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': task-1024948, 'name': Rename_Task, 'duration_secs': 0.316114} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.209167] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1186.209644] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e83ed64e-c49f-486a-bb9f-dc60cdb4d0ce {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.215638] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Waiting for the task: (returnval){ [ 1186.215638] env[63372]: value = "task-1024949" [ 1186.215638] env[63372]: _type = "Task" [ 1186.215638] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.227397] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': task-1024949, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.277179] env[63372]: DEBUG nova.compute.manager [req-5f97aeb2-60d8-4a92-8929-7b9290f66ade req-d3a8bcb6-f15a-4306-9cef-544931b6d921 service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Received event network-vif-deleted-c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1186.277374] env[63372]: INFO nova.compute.manager [req-5f97aeb2-60d8-4a92-8929-7b9290f66ade req-d3a8bcb6-f15a-4306-9cef-544931b6d921 service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Neutron deleted interface c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8; detaching it from the instance and deleting it from the info cache [ 1186.277549] env[63372]: DEBUG nova.network.neutron [req-5f97aeb2-60d8-4a92-8929-7b9290f66ade req-d3a8bcb6-f15a-4306-9cef-544931b6d921 service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.725749] env[63372]: DEBUG oslo_vmware.api [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': task-1024949, 'name': PowerOnVM_Task, 'duration_secs': 0.45802} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.726070] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1186.726219] env[63372]: INFO nova.compute.manager [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Took 6.63 seconds to spawn the instance on the hypervisor. [ 1186.726483] env[63372]: DEBUG nova.compute.manager [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1186.727255] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a37112-0cef-4b47-8b21-4e97992b2f7c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.755077] env[63372]: DEBUG nova.network.neutron [-] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.781215] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f99c7913-e6fb-470b-bd96-de4871ebd325 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.790969] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e33b0ee6-3dee-4900-bb34-564159221c96 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.817205] env[63372]: DEBUG nova.compute.manager [req-5f97aeb2-60d8-4a92-8929-7b9290f66ade req-d3a8bcb6-f15a-4306-9cef-544931b6d921 service nova] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Detach interface failed, port_id=c9c45d69-e9d7-44a7-8bd8-41dfcbbabfb8, reason: Instance 1e7590de-55ae-4725-be5a-e3435d98151a could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1187.242812] env[63372]: INFO nova.compute.manager [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Took 11.32 seconds to build instance. [ 1187.258146] env[63372]: INFO nova.compute.manager [-] [instance: 1e7590de-55ae-4725-be5a-e3435d98151a] Took 1.43 seconds to deallocate network for instance. [ 1187.744725] env[63372]: DEBUG oslo_concurrency.lockutils [None req-1f21063e-5c45-4356-b56b-1caace9bc9b0 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Lock "ea851e4f-0ccc-487c-ac4a-2d658a9337ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.831s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.764316] env[63372]: DEBUG oslo_concurrency.lockutils [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.764589] env[63372]: DEBUG oslo_concurrency.lockutils [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.764814] env[63372]: DEBUG nova.objects.instance [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lazy-loading 'resources' on Instance uuid 1e7590de-55ae-4725-be5a-e3435d98151a {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.114860] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Acquiring lock "ea851e4f-0ccc-487c-ac4a-2d658a9337ae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.115153] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Lock "ea851e4f-0ccc-487c-ac4a-2d658a9337ae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.115369] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Acquiring lock "ea851e4f-0ccc-487c-ac4a-2d658a9337ae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.115564] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Lock "ea851e4f-0ccc-487c-ac4a-2d658a9337ae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.115737] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Lock "ea851e4f-0ccc-487c-ac4a-2d658a9337ae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.117821] env[63372]: INFO nova.compute.manager [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Terminating instance [ 1188.119584] env[63372]: DEBUG nova.compute.manager [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1188.119781] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1188.120621] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16bb403e-e326-4f45-80bf-9ad3af772e3a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.130216] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1188.130440] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f2a211c-e1ef-4db9-a900-41f8cd409f0a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.136831] env[63372]: DEBUG oslo_vmware.api [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Waiting for the task: (returnval){ [ 1188.136831] env[63372]: value = "task-1024950" [ 1188.136831] env[63372]: _type = "Task" [ 1188.136831] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.144578] env[63372]: DEBUG oslo_vmware.api [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': task-1024950, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.332242] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81efd8a5-4e55-4079-9eb2-67a156d752f0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.339856] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a8f43e-7419-44c5-8974-122089c1933a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.368817] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4e5d93-3b37-4207-9e7f-f105c5d7aada {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.375661] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee16c0d1-a71f-4d11-95ad-17994c2b80b1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.388382] env[63372]: DEBUG nova.compute.provider_tree [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1188.646277] env[63372]: DEBUG oslo_vmware.api [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': task-1024950, 'name': PowerOffVM_Task, 'duration_secs': 0.195138} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.646548] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1188.646716] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1188.646958] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af288e9a-82ff-4812-a3eb-4f3297cbb5d8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.714841] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1188.715080] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1188.715272] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Deleting the datastore file [datastore2] ea851e4f-0ccc-487c-ac4a-2d658a9337ae {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1188.715527] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f618443-68dc-4f29-91f3-409594e84a25 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.721949] env[63372]: DEBUG oslo_vmware.api [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Waiting for the task: (returnval){ [ 1188.721949] env[63372]: value = "task-1024952" [ 1188.721949] env[63372]: _type = "Task" [ 1188.721949] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.730390] env[63372]: DEBUG oslo_vmware.api [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': task-1024952, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.908188] env[63372]: ERROR nova.scheduler.client.report [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [req-82e42ad4-6b01-4f5b-9f69-aa3d473501b8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID c1b110ca-a185-44c2-ba1c-f05b3b420add. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-82e42ad4-6b01-4f5b-9f69-aa3d473501b8"}]} [ 1188.923282] env[63372]: DEBUG nova.scheduler.client.report [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Refreshing inventories for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1188.934787] env[63372]: DEBUG nova.scheduler.client.report [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Updating ProviderTree inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1188.934981] env[63372]: DEBUG nova.compute.provider_tree [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 185, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1188.944587] env[63372]: DEBUG nova.scheduler.client.report [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Refreshing aggregate associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, aggregates: None {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1188.962519] env[63372]: DEBUG nova.scheduler.client.report [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Refreshing trait associations for resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63372) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1189.011813] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-791408e8-1a5c-467a-a299-52132c257e76 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.019434] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c12cfc46-e750-4250-a636-42034af61bbb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.049151] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c911c8-d67a-432b-a5d8-692976b57ddb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.056116] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1414781-33b2-4af9-9658-ebdc1a1bf13a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.068800] env[63372]: DEBUG nova.compute.provider_tree [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1189.231874] env[63372]: DEBUG oslo_vmware.api [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Task: {'id': task-1024952, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139897} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.232174] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1189.232374] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1189.232555] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1189.232730] env[63372]: INFO nova.compute.manager [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1189.232965] env[63372]: DEBUG oslo.service.loopingcall [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1189.233240] env[63372]: DEBUG nova.compute.manager [-] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1189.233366] env[63372]: DEBUG nova.network.neutron [-] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1189.472985] env[63372]: DEBUG nova.compute.manager [req-7f1e9629-b050-4697-b26f-508d1d838940 req-eb9ef3ff-729a-446e-8dfe-7aca1d957473 service nova] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Received event network-vif-deleted-60967e30-46f4-4313-b1dc-5df97f4eef71 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1189.473237] env[63372]: INFO nova.compute.manager [req-7f1e9629-b050-4697-b26f-508d1d838940 req-eb9ef3ff-729a-446e-8dfe-7aca1d957473 service nova] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Neutron deleted interface 60967e30-46f4-4313-b1dc-5df97f4eef71; detaching it from the instance and deleting it from the info cache [ 1189.473415] env[63372]: DEBUG nova.network.neutron [req-7f1e9629-b050-4697-b26f-508d1d838940 req-eb9ef3ff-729a-446e-8dfe-7aca1d957473 service nova] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.596975] env[63372]: DEBUG nova.scheduler.client.report [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Updated inventory for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with generation 167 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1189.597281] env[63372]: DEBUG nova.compute.provider_tree [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Updating resource provider c1b110ca-a185-44c2-ba1c-f05b3b420add generation from 167 to 168 during operation: update_inventory {{(pid=63372) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1189.597462] env[63372]: DEBUG nova.compute.provider_tree [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Updating inventory in ProviderTree for provider c1b110ca-a185-44c2-ba1c-f05b3b420add with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1189.949553] env[63372]: DEBUG nova.network.neutron [-] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.976028] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68934e82-3a64-4af0-936e-afe4f93b71ed {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.984784] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d278e6-90e8-4a3c-a0b5-550fae928aad {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.010558] env[63372]: DEBUG nova.compute.manager [req-7f1e9629-b050-4697-b26f-508d1d838940 req-eb9ef3ff-729a-446e-8dfe-7aca1d957473 service nova] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Detach interface failed, port_id=60967e30-46f4-4313-b1dc-5df97f4eef71, reason: Instance ea851e4f-0ccc-487c-ac4a-2d658a9337ae could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1190.102612] env[63372]: DEBUG oslo_concurrency.lockutils [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.338s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.122087] env[63372]: INFO nova.scheduler.client.report [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Deleted allocations for instance 1e7590de-55ae-4725-be5a-e3435d98151a [ 1190.452595] env[63372]: INFO nova.compute.manager [-] [instance: ea851e4f-0ccc-487c-ac4a-2d658a9337ae] Took 1.22 seconds to deallocate network for instance. [ 1190.629632] env[63372]: DEBUG oslo_concurrency.lockutils [None req-51eb4457-e15b-4b89-9708-03c54764a991 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "1e7590de-55ae-4725-be5a-e3435d98151a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.927s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.958945] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.959267] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.959450] env[63372]: DEBUG nova.objects.instance [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Lazy-loading 'resources' on Instance uuid ea851e4f-0ccc-487c-ac4a-2d658a9337ae {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1191.509028] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d3fbb2-d1db-48eb-bcb2-c96a920a3767 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.516639] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a31b555-261b-464d-8095-88afe39d7605 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.546927] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45fee7c3-74df-4c2b-be71-9417802531f8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.554645] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d73191-c76c-45d8-9365-eb278e2ae599 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.567370] env[63372]: DEBUG nova.compute.provider_tree [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1191.634284] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.634547] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1192.070520] env[63372]: DEBUG nova.scheduler.client.report [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1192.136746] env[63372]: INFO nova.compute.manager [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Detaching volume a7250982-c18c-4394-a135-fdca2a2eb278 [ 1192.167459] env[63372]: INFO nova.virt.block_device [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Attempting to driver detach volume a7250982-c18c-4394-a135-fdca2a2eb278 from mountpoint /dev/sdb [ 1192.167713] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Volume detach. Driver type: vmdk {{(pid=63372) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1192.167901] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227474', 'volume_id': 'a7250982-c18c-4394-a135-fdca2a2eb278', 'name': 'volume-a7250982-c18c-4394-a135-fdca2a2eb278', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e', 'attached_at': '', 'detached_at': '', 'volume_id': 'a7250982-c18c-4394-a135-fdca2a2eb278', 'serial': 'a7250982-c18c-4394-a135-fdca2a2eb278'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1192.168803] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9810e1ba-c510-43fc-a698-ccae6c192642 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.190129] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2459bf-5bb1-4db0-8f37-2f9669ac6c25 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.196874] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6abc4602-814c-403d-ab03-ba9d5780f753 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.216386] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe012fb2-bcff-4948-814a-f1b6e6947855 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.230117] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] The volume has not been displaced from its original location: [datastore1] volume-a7250982-c18c-4394-a135-fdca2a2eb278/volume-a7250982-c18c-4394-a135-fdca2a2eb278.vmdk. No consolidation needed. {{(pid=63372) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1192.235378] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1192.235635] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-352a88f2-60ca-4332-9191-7290db1e1df7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.252334] env[63372]: DEBUG oslo_vmware.api [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1192.252334] env[63372]: value = "task-1024953" [ 1192.252334] env[63372]: _type = "Task" [ 1192.252334] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.260859] env[63372]: DEBUG oslo_vmware.api [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024953, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.576106] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.616s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1192.593185] env[63372]: INFO nova.scheduler.client.report [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Deleted allocations for instance ea851e4f-0ccc-487c-ac4a-2d658a9337ae [ 1192.761683] env[63372]: DEBUG oslo_vmware.api [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024953, 'name': ReconfigVM_Task, 'duration_secs': 0.230143} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.762687] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1192.767122] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91bbd9a0-304e-4973-b3dc-9df1a1bbd00f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.781636] env[63372]: DEBUG oslo_vmware.api [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1192.781636] env[63372]: value = "task-1024954" [ 1192.781636] env[63372]: _type = "Task" [ 1192.781636] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.790406] env[63372]: DEBUG oslo_vmware.api [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024954, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.100562] env[63372]: DEBUG oslo_concurrency.lockutils [None req-ebd33d16-457e-427f-a08a-68b7d6f3a594 tempest-ServerAddressesTestJSON-1942159525 tempest-ServerAddressesTestJSON-1942159525-project-member] Lock "ea851e4f-0ccc-487c-ac4a-2d658a9337ae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.985s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.291514] env[63372]: DEBUG oslo_vmware.api [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024954, 'name': ReconfigVM_Task, 'duration_secs': 0.134485} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.291841] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227474', 'volume_id': 'a7250982-c18c-4394-a135-fdca2a2eb278', 'name': 'volume-a7250982-c18c-4394-a135-fdca2a2eb278', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e', 'attached_at': '', 'detached_at': '', 'volume_id': 'a7250982-c18c-4394-a135-fdca2a2eb278', 'serial': 'a7250982-c18c-4394-a135-fdca2a2eb278'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1193.840808] env[63372]: DEBUG nova.objects.instance [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lazy-loading 'flavor' on Instance uuid 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1194.853638] env[63372]: DEBUG oslo_concurrency.lockutils [None req-c083b9f3-19c1-4010-b100-246d5cf81094 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.219s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.894324] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.894693] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.894930] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.895186] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.895408] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.897455] env[63372]: INFO nova.compute.manager [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Terminating instance [ 1195.899201] env[63372]: DEBUG nova.compute.manager [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1195.899438] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1195.900306] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8458685d-643e-4004-9a56-3c828944ae9c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.908227] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1195.908457] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b76922f1-09e3-4dda-a0f8-69f2e68ed0be {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.914725] env[63372]: DEBUG oslo_vmware.api [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1195.914725] env[63372]: value = "task-1024955" [ 1195.914725] env[63372]: _type = "Task" [ 1195.914725] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.921961] env[63372]: DEBUG oslo_vmware.api [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024955, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.424923] env[63372]: DEBUG oslo_vmware.api [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024955, 'name': PowerOffVM_Task, 'duration_secs': 0.167482} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.425216] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1196.425388] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1196.425623] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d97a410-5c7b-43de-8393-7cd15b66f507 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.493753] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1196.493969] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1196.494173] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Deleting the datastore file [datastore1] 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1196.494425] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d528648a-2000-4a04-b5b8-caee37cf764d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.500600] env[63372]: DEBUG oslo_vmware.api [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1196.500600] env[63372]: value = "task-1024957" [ 1196.500600] env[63372]: _type = "Task" [ 1196.500600] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.507590] env[63372]: DEBUG oslo_vmware.api [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024957, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.010511] env[63372]: DEBUG oslo_vmware.api [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024957, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144951} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.010890] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1197.010939] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1197.011134] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1197.011316] env[63372]: INFO nova.compute.manager [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1197.011558] env[63372]: DEBUG oslo.service.loopingcall [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1197.011749] env[63372]: DEBUG nova.compute.manager [-] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1197.011844] env[63372]: DEBUG nova.network.neutron [-] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1197.450569] env[63372]: DEBUG nova.compute.manager [req-f7e8abb6-e452-417e-888b-6e218c8754e5 req-e2053e70-1130-4015-a121-c2e3f11e3062 service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Received event network-vif-deleted-ba2bc8da-ac45-4f6f-a227-c425f0b2d76d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1197.450963] env[63372]: INFO nova.compute.manager [req-f7e8abb6-e452-417e-888b-6e218c8754e5 req-e2053e70-1130-4015-a121-c2e3f11e3062 service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Neutron deleted interface ba2bc8da-ac45-4f6f-a227-c425f0b2d76d; detaching it from the instance and deleting it from the info cache [ 1197.450963] env[63372]: DEBUG nova.network.neutron [req-f7e8abb6-e452-417e-888b-6e218c8754e5 req-e2053e70-1130-4015-a121-c2e3f11e3062 service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.192215] env[63372]: DEBUG nova.network.neutron [-] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.194308] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-92a312e1-9652-444e-9d8e-7e1a5412eb2e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.204639] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5caf90cc-af5b-4b22-bfac-b729b70f4349 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.229687] env[63372]: DEBUG nova.compute.manager [req-f7e8abb6-e452-417e-888b-6e218c8754e5 req-e2053e70-1130-4015-a121-c2e3f11e3062 service nova] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Detach interface failed, port_id=ba2bc8da-ac45-4f6f-a227-c425f0b2d76d, reason: Instance 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1198.694485] env[63372]: INFO nova.compute.manager [-] [instance: 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e] Took 1.68 seconds to deallocate network for instance. [ 1199.200867] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1199.201248] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1199.201388] env[63372]: DEBUG nova.objects.instance [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lazy-loading 'resources' on Instance uuid 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1199.744203] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a53f07e-cf41-4ec8-badd-65ef54f50b55 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.751646] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0682886-0658-453b-85e5-8de7a0cb8420 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.781317] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe6b52c-b5fe-43d0-b7b5-95f336f14ea2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.788216] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-746aa2cb-30ee-41af-9d84-f12ec35f9067 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.802113] env[63372]: DEBUG nova.compute.provider_tree [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1200.305267] env[63372]: DEBUG nova.scheduler.client.report [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1200.811060] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.609s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1200.828837] env[63372]: INFO nova.scheduler.client.report [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Deleted allocations for instance 9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e [ 1201.337258] env[63372]: DEBUG oslo_concurrency.lockutils [None req-3012ae6f-1b9a-45c8-afce-aed7e6aabe26 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "9d7bfb0c-65f4-4dd7-8998-b5114a1e4f5e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.442s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1201.894689] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.894962] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.895136] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Starting heal instance info cache {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1203.853412] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.853671] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1204.356024] env[63372]: DEBUG nova.compute.manager [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1204.877547] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.877826] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1204.879351] env[63372]: INFO nova.compute.claims [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1204.908851] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Didn't find any instances for network info cache update. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1204.909069] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1204.909237] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1204.909388] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1204.909536] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1204.909675] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1204.909819] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1204.909943] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63372) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1205.925184] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763154b7-ffae-473d-9c36-fa2bb3820b2b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.932647] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c2e1d9-451c-48c2-8368-29c88e586ade {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.963128] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458be73b-f1e9-402b-9d7b-2e48aef5368c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.970926] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac0e609f-1b0b-4051-a462-2e604e1186a4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.984628] env[63372]: DEBUG nova.compute.provider_tree [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1206.128013] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager.update_available_resource {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1206.487870] env[63372]: DEBUG nova.scheduler.client.report [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1206.631375] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1206.992642] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.115s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1206.993204] env[63372]: DEBUG nova.compute.manager [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1206.996397] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.365s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1206.996620] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1206.996814] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63372) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1206.997714] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246d77d8-0427-4259-9303-d3f1b7ac6fc7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.007781] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ed2d0d-950a-493a-a893-a07f8b5a8146 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.023064] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6966c2ca-b38f-48b4-a04e-d2d8f4f2ce36 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.030334] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ca855a-1eb3-4e9b-a879-5a9034433106 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.059914] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180843MB free_disk=186GB free_vcpus=48 pci_devices=None {{(pid=63372) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1207.060066] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.060260] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.505056] env[63372]: DEBUG nova.compute.utils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1207.505881] env[63372]: DEBUG nova.compute.manager [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1207.506482] env[63372]: DEBUG nova.network.neutron [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1207.553339] env[63372]: DEBUG nova.policy [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc5c59e1419843ad90b1bc33f0a55c31', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8452e71bbbd04222bb6b868017b1b6ed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1207.830920] env[63372]: DEBUG nova.network.neutron [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Successfully created port: 77eef7fd-df07-4cd7-9a4c-0ac51961a494 {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1208.009963] env[63372]: DEBUG nova.compute.manager [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1208.080739] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance fafed45e-c9dc-4196-883f-a23570fd0583 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1208.080919] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance e922c0c6-50fa-4a4e-9961-bc5bc3a880e7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1208.081109] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1208.081251] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1208.113381] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4218e95-b4d8-436a-aa26-f65bb83b407f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.121127] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a35e15a-23f0-4b08-b85c-42e65c609d43 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.150012] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a3ea89-0fbf-418a-a043-54256bd216d9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.158011] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9234f67f-6496-44e3-927a-39b5e5d4a44d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.171134] env[63372]: DEBUG nova.compute.provider_tree [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1208.397532] env[63372]: DEBUG oslo_concurrency.lockutils [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "fafed45e-c9dc-4196-883f-a23570fd0583" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.397532] env[63372]: DEBUG oslo_concurrency.lockutils [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.675049] env[63372]: DEBUG nova.scheduler.client.report [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1208.900772] env[63372]: INFO nova.compute.manager [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Detaching volume 598ca7c3-f9fb-4fe4-939a-9ca37fc19323 [ 1208.937048] env[63372]: INFO nova.virt.block_device [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Attempting to driver detach volume 598ca7c3-f9fb-4fe4-939a-9ca37fc19323 from mountpoint /dev/sdb [ 1208.937370] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Volume detach. Driver type: vmdk {{(pid=63372) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1208.937511] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227479', 'volume_id': '598ca7c3-f9fb-4fe4-939a-9ca37fc19323', 'name': 'volume-598ca7c3-f9fb-4fe4-939a-9ca37fc19323', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'fafed45e-c9dc-4196-883f-a23570fd0583', 'attached_at': '', 'detached_at': '', 'volume_id': '598ca7c3-f9fb-4fe4-939a-9ca37fc19323', 'serial': '598ca7c3-f9fb-4fe4-939a-9ca37fc19323'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1208.938410] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2fd46e0-eb9c-4ead-a1cd-5f5d6023215a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.960487] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62ebc88-aea7-4d99-9619-bd55290453d2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.967379] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a17ece8-46a5-4126-9a7d-07d24f191cb8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.986960] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e4b3a4-a2d0-432d-9ff8-af03ea88a881 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.001220] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] The volume has not been displaced from its original location: [datastore1] volume-598ca7c3-f9fb-4fe4-939a-9ca37fc19323/volume-598ca7c3-f9fb-4fe4-939a-9ca37fc19323.vmdk. No consolidation needed. {{(pid=63372) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1209.006439] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Reconfiguring VM instance instance-0000006f to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1209.006709] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f3f6371-c657-4b60-9f69-bac14269b117 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.022348] env[63372]: DEBUG nova.compute.manager [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1209.025259] env[63372]: DEBUG oslo_vmware.api [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1209.025259] env[63372]: value = "task-1024958" [ 1209.025259] env[63372]: _type = "Task" [ 1209.025259] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.032769] env[63372]: DEBUG oslo_vmware.api [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024958, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.047918] env[63372]: DEBUG nova.virt.hardware [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1209.048198] env[63372]: DEBUG nova.virt.hardware [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1209.048377] env[63372]: DEBUG nova.virt.hardware [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1209.048594] env[63372]: DEBUG nova.virt.hardware [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1209.048746] env[63372]: DEBUG nova.virt.hardware [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1209.048907] env[63372]: DEBUG nova.virt.hardware [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1209.049152] env[63372]: DEBUG nova.virt.hardware [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1209.049334] env[63372]: DEBUG nova.virt.hardware [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1209.049511] env[63372]: DEBUG nova.virt.hardware [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1209.049677] env[63372]: DEBUG nova.virt.hardware [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1209.049852] env[63372]: DEBUG nova.virt.hardware [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1209.050645] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8e2b89c-1e52-4d92-b3dc-4f4fce003b59 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.057944] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0119fa6-e035-4a9d-9f55-02e148a6c780 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.178822] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1209.179031] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.119s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.212858] env[63372]: DEBUG nova.compute.manager [req-e54d454f-b8db-46c3-ba69-64f6f3016152 req-22d4df0a-0a6b-46b0-8f62-0510b45ae00c service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Received event network-vif-plugged-77eef7fd-df07-4cd7-9a4c-0ac51961a494 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1209.212858] env[63372]: DEBUG oslo_concurrency.lockutils [req-e54d454f-b8db-46c3-ba69-64f6f3016152 req-22d4df0a-0a6b-46b0-8f62-0510b45ae00c service nova] Acquiring lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.212858] env[63372]: DEBUG oslo_concurrency.lockutils [req-e54d454f-b8db-46c3-ba69-64f6f3016152 req-22d4df0a-0a6b-46b0-8f62-0510b45ae00c service nova] Lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.213174] env[63372]: DEBUG oslo_concurrency.lockutils [req-e54d454f-b8db-46c3-ba69-64f6f3016152 req-22d4df0a-0a6b-46b0-8f62-0510b45ae00c service nova] Lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.213326] env[63372]: DEBUG nova.compute.manager [req-e54d454f-b8db-46c3-ba69-64f6f3016152 req-22d4df0a-0a6b-46b0-8f62-0510b45ae00c service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] No waiting events found dispatching network-vif-plugged-77eef7fd-df07-4cd7-9a4c-0ac51961a494 {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1209.213540] env[63372]: WARNING nova.compute.manager [req-e54d454f-b8db-46c3-ba69-64f6f3016152 req-22d4df0a-0a6b-46b0-8f62-0510b45ae00c service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Received unexpected event network-vif-plugged-77eef7fd-df07-4cd7-9a4c-0ac51961a494 for instance with vm_state building and task_state spawning. [ 1209.296249] env[63372]: DEBUG nova.network.neutron [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Successfully updated port: 77eef7fd-df07-4cd7-9a4c-0ac51961a494 {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1209.537324] env[63372]: DEBUG oslo_vmware.api [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024958, 'name': ReconfigVM_Task, 'duration_secs': 0.217027} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.537612] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Reconfigured VM instance instance-0000006f to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1209.542127] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d036081-9a45-46ba-a2ca-a62bd9b0743c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.556894] env[63372]: DEBUG oslo_vmware.api [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1209.556894] env[63372]: value = "task-1024959" [ 1209.556894] env[63372]: _type = "Task" [ 1209.556894] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.564194] env[63372]: DEBUG oslo_vmware.api [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024959, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.799147] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "refresh_cache-e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1209.799257] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired lock "refresh_cache-e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1209.799405] env[63372]: DEBUG nova.network.neutron [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1210.067853] env[63372]: DEBUG oslo_vmware.api [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024959, 'name': ReconfigVM_Task, 'duration_secs': 0.129988} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.068407] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227479', 'volume_id': '598ca7c3-f9fb-4fe4-939a-9ca37fc19323', 'name': 'volume-598ca7c3-f9fb-4fe4-939a-9ca37fc19323', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'fafed45e-c9dc-4196-883f-a23570fd0583', 'attached_at': '', 'detached_at': '', 'volume_id': '598ca7c3-f9fb-4fe4-939a-9ca37fc19323', 'serial': '598ca7c3-f9fb-4fe4-939a-9ca37fc19323'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1210.330230] env[63372]: DEBUG nova.network.neutron [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1210.450963] env[63372]: DEBUG nova.network.neutron [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Updating instance_info_cache with network_info: [{"id": "77eef7fd-df07-4cd7-9a4c-0ac51961a494", "address": "fa:16:3e:f3:8f:fc", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77eef7fd-df", "ovs_interfaceid": "77eef7fd-df07-4cd7-9a4c-0ac51961a494", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1210.607306] env[63372]: DEBUG nova.objects.instance [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lazy-loading 'flavor' on Instance uuid fafed45e-c9dc-4196-883f-a23570fd0583 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1210.953415] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Releasing lock "refresh_cache-e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1210.953748] env[63372]: DEBUG nova.compute.manager [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Instance network_info: |[{"id": "77eef7fd-df07-4cd7-9a4c-0ac51961a494", "address": "fa:16:3e:f3:8f:fc", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77eef7fd-df", "ovs_interfaceid": "77eef7fd-df07-4cd7-9a4c-0ac51961a494", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1210.954175] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:8f:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '77eef7fd-df07-4cd7-9a4c-0ac51961a494', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1210.961670] env[63372]: DEBUG oslo.service.loopingcall [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1210.961881] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1210.962117] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fad328ef-f4fc-46cd-b4ab-fa9e1e282d96 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.982142] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1210.982142] env[63372]: value = "task-1024960" [ 1210.982142] env[63372]: _type = "Task" [ 1210.982142] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.989797] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024960, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.237717] env[63372]: DEBUG nova.compute.manager [req-e7ebd65c-6129-46ee-bf15-20f4451e3e47 req-a9351902-25ef-4dda-9217-e90f463afd2e service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Received event network-changed-77eef7fd-df07-4cd7-9a4c-0ac51961a494 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1211.238073] env[63372]: DEBUG nova.compute.manager [req-e7ebd65c-6129-46ee-bf15-20f4451e3e47 req-a9351902-25ef-4dda-9217-e90f463afd2e service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Refreshing instance network info cache due to event network-changed-77eef7fd-df07-4cd7-9a4c-0ac51961a494. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1211.238182] env[63372]: DEBUG oslo_concurrency.lockutils [req-e7ebd65c-6129-46ee-bf15-20f4451e3e47 req-a9351902-25ef-4dda-9217-e90f463afd2e service nova] Acquiring lock "refresh_cache-e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1211.238337] env[63372]: DEBUG oslo_concurrency.lockutils [req-e7ebd65c-6129-46ee-bf15-20f4451e3e47 req-a9351902-25ef-4dda-9217-e90f463afd2e service nova] Acquired lock "refresh_cache-e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1211.238510] env[63372]: DEBUG nova.network.neutron [req-e7ebd65c-6129-46ee-bf15-20f4451e3e47 req-a9351902-25ef-4dda-9217-e90f463afd2e service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Refreshing network info cache for port 77eef7fd-df07-4cd7-9a4c-0ac51961a494 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1211.491891] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024960, 'name': CreateVM_Task, 'duration_secs': 0.294036} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.492099] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1211.492824] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1211.492993] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1211.493333] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1211.493575] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6d3fdeb-b204-426f-bfa7-0afd9c5fc9a6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.498226] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1211.498226] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52811609-23e3-b2ab-8145-9a1b7774adb9" [ 1211.498226] env[63372]: _type = "Task" [ 1211.498226] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.506262] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52811609-23e3-b2ab-8145-9a1b7774adb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.563820] env[63372]: DEBUG oslo_concurrency.lockutils [None req-599f66f9-1906-4087-8973-2bc3dcb052c3 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "fafed45e-c9dc-4196-883f-a23570fd0583" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1211.613466] env[63372]: DEBUG oslo_concurrency.lockutils [None req-380cbc0e-52a5-417d-b6d4-25012eedbeff tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.216s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1211.614588] env[63372]: DEBUG oslo_concurrency.lockutils [None req-599f66f9-1906-4087-8973-2bc3dcb052c3 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.051s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1211.614797] env[63372]: DEBUG nova.compute.manager [None req-599f66f9-1906-4087-8973-2bc3dcb052c3 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1211.615803] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c44fe0ce-3e27-4540-8a68-99e59e36746b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.622126] env[63372]: DEBUG nova.compute.manager [None req-599f66f9-1906-4087-8973-2bc3dcb052c3 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63372) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1211.622723] env[63372]: DEBUG nova.objects.instance [None req-599f66f9-1906-4087-8973-2bc3dcb052c3 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lazy-loading 'flavor' on Instance uuid fafed45e-c9dc-4196-883f-a23570fd0583 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1211.933261] env[63372]: DEBUG nova.network.neutron [req-e7ebd65c-6129-46ee-bf15-20f4451e3e47 req-a9351902-25ef-4dda-9217-e90f463afd2e service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Updated VIF entry in instance network info cache for port 77eef7fd-df07-4cd7-9a4c-0ac51961a494. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1211.933682] env[63372]: DEBUG nova.network.neutron [req-e7ebd65c-6129-46ee-bf15-20f4451e3e47 req-a9351902-25ef-4dda-9217-e90f463afd2e service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Updating instance_info_cache with network_info: [{"id": "77eef7fd-df07-4cd7-9a4c-0ac51961a494", "address": "fa:16:3e:f3:8f:fc", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77eef7fd-df", "ovs_interfaceid": "77eef7fd-df07-4cd7-9a4c-0ac51961a494", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.008978] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52811609-23e3-b2ab-8145-9a1b7774adb9, 'name': SearchDatastore_Task, 'duration_secs': 0.012323} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.009308] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1212.009591] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1212.010029] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1212.010029] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.010199] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1212.010412] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af19bbc1-96d6-4ff9-be0e-bfb90796c61a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.018609] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1212.018782] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1212.019491] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70fa8b89-68c8-47bc-9142-69caf801bbf8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.024567] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1212.024567] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]527c5ada-5ed3-a719-890e-8656eaf52ad1" [ 1212.024567] env[63372]: _type = "Task" [ 1212.024567] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.031485] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527c5ada-5ed3-a719-890e-8656eaf52ad1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.127076] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-599f66f9-1906-4087-8973-2bc3dcb052c3 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1212.127360] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df8919e3-1fb3-492b-9ad6-db42fc95f1dc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.134842] env[63372]: DEBUG oslo_vmware.api [None req-599f66f9-1906-4087-8973-2bc3dcb052c3 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1212.134842] env[63372]: value = "task-1024961" [ 1212.134842] env[63372]: _type = "Task" [ 1212.134842] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.142597] env[63372]: DEBUG oslo_vmware.api [None req-599f66f9-1906-4087-8973-2bc3dcb052c3 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024961, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.174431] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.436663] env[63372]: DEBUG oslo_concurrency.lockutils [req-e7ebd65c-6129-46ee-bf15-20f4451e3e47 req-a9351902-25ef-4dda-9217-e90f463afd2e service nova] Releasing lock "refresh_cache-e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1212.535370] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]527c5ada-5ed3-a719-890e-8656eaf52ad1, 'name': SearchDatastore_Task, 'duration_secs': 0.008323} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.536101] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f98bac2-8ea2-49c4-802b-e33c757b8252 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.540886] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1212.540886] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52fa3a52-22eb-ecd6-b9a7-11d5a7dbd2e1" [ 1212.540886] env[63372]: _type = "Task" [ 1212.540886] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.548269] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fa3a52-22eb-ecd6-b9a7-11d5a7dbd2e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.644064] env[63372]: DEBUG oslo_vmware.api [None req-599f66f9-1906-4087-8973-2bc3dcb052c3 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024961, 'name': PowerOffVM_Task, 'duration_secs': 0.202887} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.644328] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-599f66f9-1906-4087-8973-2bc3dcb052c3 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1212.644508] env[63372]: DEBUG nova.compute.manager [None req-599f66f9-1906-4087-8973-2bc3dcb052c3 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1212.645263] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a0bd0f-9703-4db2-ad3b-27cd6a54a052 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.051920] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52fa3a52-22eb-ecd6-b9a7-11d5a7dbd2e1, 'name': SearchDatastore_Task, 'duration_secs': 0.009127} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.052192] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1213.052479] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] e922c0c6-50fa-4a4e-9961-bc5bc3a880e7/e922c0c6-50fa-4a4e-9961-bc5bc3a880e7.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1213.052728] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d2405c0-83a0-4451-82d2-c1a9f2115222 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.058884] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1213.058884] env[63372]: value = "task-1024962" [ 1213.058884] env[63372]: _type = "Task" [ 1213.058884] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.066264] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024962, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.156598] env[63372]: DEBUG oslo_concurrency.lockutils [None req-599f66f9-1906-4087-8973-2bc3dcb052c3 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.542s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.569859] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024962, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.432438} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.570142] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore1] e922c0c6-50fa-4a4e-9961-bc5bc3a880e7/e922c0c6-50fa-4a4e-9961-bc5bc3a880e7.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1213.570297] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1213.570548] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-470b8eed-6fea-42dc-bcfc-eeea2c113ea9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.577159] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1213.577159] env[63372]: value = "task-1024963" [ 1213.577159] env[63372]: _type = "Task" [ 1213.577159] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.585301] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024963, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.985982] env[63372]: DEBUG nova.objects.instance [None req-68fe89f8-722b-4c29-875a-e25c93858ba8 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lazy-loading 'flavor' on Instance uuid fafed45e-c9dc-4196-883f-a23570fd0583 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1214.087045] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024963, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06081} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.087373] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1214.089224] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b7c8e4-b824-4a51-a236-a4c541ac2f0b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.110137] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] e922c0c6-50fa-4a4e-9961-bc5bc3a880e7/e922c0c6-50fa-4a4e-9961-bc5bc3a880e7.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1214.110397] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44480616-c081-49b7-8ca6-0091e3d134bf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.130138] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1214.130138] env[63372]: value = "task-1024964" [ 1214.130138] env[63372]: _type = "Task" [ 1214.130138] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.138222] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024964, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.491821] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68fe89f8-722b-4c29-875a-e25c93858ba8 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1214.492094] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68fe89f8-722b-4c29-875a-e25c93858ba8 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquired lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1214.492209] env[63372]: DEBUG nova.network.neutron [None req-68fe89f8-722b-4c29-875a-e25c93858ba8 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1214.492437] env[63372]: DEBUG nova.objects.instance [None req-68fe89f8-722b-4c29-875a-e25c93858ba8 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lazy-loading 'info_cache' on Instance uuid fafed45e-c9dc-4196-883f-a23570fd0583 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1214.641066] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024964, 'name': ReconfigVM_Task, 'duration_secs': 0.348443} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.641066] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Reconfigured VM instance instance-00000072 to attach disk [datastore1] e922c0c6-50fa-4a4e-9961-bc5bc3a880e7/e922c0c6-50fa-4a4e-9961-bc5bc3a880e7.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1214.641685] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c1bf2aa-07ba-4469-9a61-dcdc3c1b3d3d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.648936] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1214.648936] env[63372]: value = "task-1024965" [ 1214.648936] env[63372]: _type = "Task" [ 1214.648936] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.656502] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024965, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.996164] env[63372]: DEBUG nova.objects.base [None req-68fe89f8-722b-4c29-875a-e25c93858ba8 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63372) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1215.158628] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024965, 'name': Rename_Task, 'duration_secs': 0.13695} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.158918] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1215.159179] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ac25383-d7ba-4417-bc34-781c5fe038b1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.165979] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1215.165979] env[63372]: value = "task-1024966" [ 1215.165979] env[63372]: _type = "Task" [ 1215.165979] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.173330] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024966, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.674959] env[63372]: DEBUG oslo_vmware.api [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024966, 'name': PowerOnVM_Task, 'duration_secs': 0.424449} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.675306] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1215.675438] env[63372]: INFO nova.compute.manager [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Took 6.65 seconds to spawn the instance on the hypervisor. [ 1215.675618] env[63372]: DEBUG nova.compute.manager [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1215.676361] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d42f3d3-31ff-4e0d-b19f-39dbe29b815c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.720660] env[63372]: DEBUG nova.network.neutron [None req-68fe89f8-722b-4c29-875a-e25c93858ba8 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Updating instance_info_cache with network_info: [{"id": "7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db", "address": "fa:16:3e:94:43:d7", "network": {"id": "b161f533-292b-431a-a171-e9d61bc90105", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1928954423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4b088060a2c4159a62e40e7e627076a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ca80379-c6", "ovs_interfaceid": "7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1216.193378] env[63372]: INFO nova.compute.manager [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Took 11.33 seconds to build instance. [ 1216.223860] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68fe89f8-722b-4c29-875a-e25c93858ba8 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Releasing lock "refresh_cache-fafed45e-c9dc-4196-883f-a23570fd0583" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1216.695220] env[63372]: DEBUG oslo_concurrency.lockutils [None req-28e5cd80-b471-4c53-8196-940ac552f805 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.841s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.728491] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-68fe89f8-722b-4c29-875a-e25c93858ba8 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1216.728961] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bbe25fcb-531b-464f-92eb-007631124a54 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.736954] env[63372]: DEBUG oslo_vmware.api [None req-68fe89f8-722b-4c29-875a-e25c93858ba8 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1216.736954] env[63372]: value = "task-1024967" [ 1216.736954] env[63372]: _type = "Task" [ 1216.736954] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.747386] env[63372]: DEBUG oslo_vmware.api [None req-68fe89f8-722b-4c29-875a-e25c93858ba8 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024967, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.128096] env[63372]: DEBUG nova.compute.manager [req-91c7115a-65c6-4816-a15c-08df2d956c2f req-e60098eb-39c2-4e17-9e2d-bad27f7c9f4a service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Received event network-changed-77eef7fd-df07-4cd7-9a4c-0ac51961a494 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1217.128304] env[63372]: DEBUG nova.compute.manager [req-91c7115a-65c6-4816-a15c-08df2d956c2f req-e60098eb-39c2-4e17-9e2d-bad27f7c9f4a service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Refreshing instance network info cache due to event network-changed-77eef7fd-df07-4cd7-9a4c-0ac51961a494. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1217.128533] env[63372]: DEBUG oslo_concurrency.lockutils [req-91c7115a-65c6-4816-a15c-08df2d956c2f req-e60098eb-39c2-4e17-9e2d-bad27f7c9f4a service nova] Acquiring lock "refresh_cache-e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1217.128680] env[63372]: DEBUG oslo_concurrency.lockutils [req-91c7115a-65c6-4816-a15c-08df2d956c2f req-e60098eb-39c2-4e17-9e2d-bad27f7c9f4a service nova] Acquired lock "refresh_cache-e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.128847] env[63372]: DEBUG nova.network.neutron [req-91c7115a-65c6-4816-a15c-08df2d956c2f req-e60098eb-39c2-4e17-9e2d-bad27f7c9f4a service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Refreshing network info cache for port 77eef7fd-df07-4cd7-9a4c-0ac51961a494 {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1217.247033] env[63372]: DEBUG oslo_vmware.api [None req-68fe89f8-722b-4c29-875a-e25c93858ba8 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024967, 'name': PowerOnVM_Task, 'duration_secs': 0.390687} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.247187] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-68fe89f8-722b-4c29-875a-e25c93858ba8 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1217.247384] env[63372]: DEBUG nova.compute.manager [None req-68fe89f8-722b-4c29-875a-e25c93858ba8 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1217.248088] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a324fbd0-4fec-4e91-be40-7a498f057882 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.832946] env[63372]: DEBUG nova.network.neutron [req-91c7115a-65c6-4816-a15c-08df2d956c2f req-e60098eb-39c2-4e17-9e2d-bad27f7c9f4a service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Updated VIF entry in instance network info cache for port 77eef7fd-df07-4cd7-9a4c-0ac51961a494. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1217.833353] env[63372]: DEBUG nova.network.neutron [req-91c7115a-65c6-4816-a15c-08df2d956c2f req-e60098eb-39c2-4e17-9e2d-bad27f7c9f4a service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Updating instance_info_cache with network_info: [{"id": "77eef7fd-df07-4cd7-9a4c-0ac51961a494", "address": "fa:16:3e:f3:8f:fc", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77eef7fd-df", "ovs_interfaceid": "77eef7fd-df07-4cd7-9a4c-0ac51961a494", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1218.335839] env[63372]: DEBUG oslo_concurrency.lockutils [req-91c7115a-65c6-4816-a15c-08df2d956c2f req-e60098eb-39c2-4e17-9e2d-bad27f7c9f4a service nova] Releasing lock "refresh_cache-e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1254.428515] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "fafed45e-c9dc-4196-883f-a23570fd0583" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.428957] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.429075] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "fafed45e-c9dc-4196-883f-a23570fd0583-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.429279] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.429457] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1254.433182] env[63372]: INFO nova.compute.manager [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Terminating instance [ 1254.435142] env[63372]: DEBUG nova.compute.manager [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1254.435345] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1254.436227] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fde505-1e68-48f3-8437-2c0b433d691f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.444453] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1254.444679] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b921e9d-cd09-4665-8686-ff0a944a8230 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.451624] env[63372]: DEBUG oslo_vmware.api [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1254.451624] env[63372]: value = "task-1024968" [ 1254.451624] env[63372]: _type = "Task" [ 1254.451624] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.460673] env[63372]: DEBUG oslo_vmware.api [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024968, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.961351] env[63372]: DEBUG oslo_vmware.api [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024968, 'name': PowerOffVM_Task, 'duration_secs': 0.169232} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.961627] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1254.961798] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1254.962047] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4030bd50-46aa-4159-9760-da974be24281 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.023970] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1255.024388] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1255.068278] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1255.068525] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1255.068673] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Deleting the datastore file [datastore2] fafed45e-c9dc-4196-883f-a23570fd0583 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1255.068937] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d49ba6d-bde4-43b7-803c-d48ef5461625 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.075485] env[63372]: DEBUG oslo_vmware.api [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1255.075485] env[63372]: value = "task-1024970" [ 1255.075485] env[63372]: _type = "Task" [ 1255.075485] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.083158] env[63372]: DEBUG oslo_vmware.api [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024970, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.528288] env[63372]: DEBUG nova.compute.utils [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1255.585248] env[63372]: DEBUG oslo_vmware.api [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024970, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146498} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.585512] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1255.585699] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1255.585878] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1255.586058] env[63372]: INFO nova.compute.manager [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1255.586316] env[63372]: DEBUG oslo.service.loopingcall [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1255.586505] env[63372]: DEBUG nova.compute.manager [-] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1255.586595] env[63372]: DEBUG nova.network.neutron [-] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1256.026834] env[63372]: DEBUG nova.compute.manager [req-fad1033b-b1c6-4800-9773-498386109587 req-59a85b86-2622-4d5b-b5ec-48c8c965b32b service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Received event network-vif-deleted-7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1256.027077] env[63372]: INFO nova.compute.manager [req-fad1033b-b1c6-4800-9773-498386109587 req-59a85b86-2622-4d5b-b5ec-48c8c965b32b service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Neutron deleted interface 7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db; detaching it from the instance and deleting it from the info cache [ 1256.027264] env[63372]: DEBUG nova.network.neutron [req-fad1033b-b1c6-4800-9773-498386109587 req-59a85b86-2622-4d5b-b5ec-48c8c965b32b service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1256.031189] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1256.506277] env[63372]: DEBUG nova.network.neutron [-] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1256.529943] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b67bee0-3dfb-44ee-aab3-aa6aa105280b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.539511] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-134e6dc0-b012-41f0-bfe1-2b2fa5c0e32d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.564817] env[63372]: DEBUG nova.compute.manager [req-fad1033b-b1c6-4800-9773-498386109587 req-59a85b86-2622-4d5b-b5ec-48c8c965b32b service nova] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Detach interface failed, port_id=7ca80379-c6e5-4dd7-bf9c-6ee6bb6ab6db, reason: Instance fafed45e-c9dc-4196-883f-a23570fd0583 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1257.008919] env[63372]: INFO nova.compute.manager [-] [instance: fafed45e-c9dc-4196-883f-a23570fd0583] Took 1.42 seconds to deallocate network for instance. [ 1257.093535] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1257.093890] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1257.094161] env[63372]: INFO nova.compute.manager [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Attaching volume 0e9d1ac5-508e-4508-8626-deeb9b690c91 to /dev/sdb [ 1257.123272] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1968a395-6dcf-4957-b719-f5deb16b6879 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.130380] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f18336-a1d5-490f-8de3-e811a1fe9763 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.143515] env[63372]: DEBUG nova.virt.block_device [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Updating existing volume attachment record: 46744456-e359-49e3-889b-dfbfd179fc7e {{(pid=63372) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1257.515463] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1257.515702] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1257.515916] env[63372]: DEBUG nova.objects.instance [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lazy-loading 'resources' on Instance uuid fafed45e-c9dc-4196-883f-a23570fd0583 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1258.059891] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11bfa1de-eb82-4fcc-b2f6-172a1e767c09 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.067722] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84fc5f45-3511-4a3f-982b-40eeeecc8dbd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.096734] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a346b2f5-ba5d-447a-bd02-d6a18a92fd84 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.103939] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a49a0d3-1c44-41c0-b2ec-f6f729ddced6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.118090] env[63372]: DEBUG nova.compute.provider_tree [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1258.621397] env[63372]: DEBUG nova.scheduler.client.report [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1259.126682] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.611s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1259.145819] env[63372]: INFO nova.scheduler.client.report [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Deleted allocations for instance fafed45e-c9dc-4196-883f-a23570fd0583 [ 1259.653363] env[63372]: DEBUG oslo_concurrency.lockutils [None req-f853d557-db4a-4907-b126-f77c05e80032 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "fafed45e-c9dc-4196-883f-a23570fd0583" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.224s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1261.128056] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1261.128402] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1261.128497] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1261.128688] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1261.301186] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1261.301429] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1261.687046] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Volume attach. Driver type: vmdk {{(pid=63372) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1261.687046] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227484', 'volume_id': '0e9d1ac5-508e-4508-8626-deeb9b690c91', 'name': 'volume-0e9d1ac5-508e-4508-8626-deeb9b690c91', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e922c0c6-50fa-4a4e-9961-bc5bc3a880e7', 'attached_at': '', 'detached_at': '', 'volume_id': '0e9d1ac5-508e-4508-8626-deeb9b690c91', 'serial': '0e9d1ac5-508e-4508-8626-deeb9b690c91'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1261.687763] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8caa32c-d2b2-41be-92d3-24172809628b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.704341] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3154be-7ca9-487b-8ee8-ae42f28e2abc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.728170] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] volume-0e9d1ac5-508e-4508-8626-deeb9b690c91/volume-0e9d1ac5-508e-4508-8626-deeb9b690c91.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1261.728411] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5f2eede-8c2d-4a46-9148-c455475a197b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.746054] env[63372]: DEBUG oslo_vmware.api [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1261.746054] env[63372]: value = "task-1024973" [ 1261.746054] env[63372]: _type = "Task" [ 1261.746054] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.755147] env[63372]: DEBUG oslo_vmware.api [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024973, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.804741] env[63372]: DEBUG nova.compute.manager [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1262.256049] env[63372]: DEBUG oslo_vmware.api [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024973, 'name': ReconfigVM_Task, 'duration_secs': 0.371182} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.256411] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Reconfigured VM instance instance-00000072 to attach disk [datastore2] volume-0e9d1ac5-508e-4508-8626-deeb9b690c91/volume-0e9d1ac5-508e-4508-8626-deeb9b690c91.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1262.260993] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2656834-1e90-45cd-80d9-1135ae3fab20 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.275804] env[63372]: DEBUG oslo_vmware.api [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1262.275804] env[63372]: value = "task-1024974" [ 1262.275804] env[63372]: _type = "Task" [ 1262.275804] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.283322] env[63372]: DEBUG oslo_vmware.api [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024974, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.322188] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.322446] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.323996] env[63372]: INFO nova.compute.claims [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1262.785644] env[63372]: DEBUG oslo_vmware.api [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024974, 'name': ReconfigVM_Task, 'duration_secs': 0.148388} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.786022] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227484', 'volume_id': '0e9d1ac5-508e-4508-8626-deeb9b690c91', 'name': 'volume-0e9d1ac5-508e-4508-8626-deeb9b690c91', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e922c0c6-50fa-4a4e-9961-bc5bc3a880e7', 'attached_at': '', 'detached_at': '', 'volume_id': '0e9d1ac5-508e-4508-8626-deeb9b690c91', 'serial': '0e9d1ac5-508e-4508-8626-deeb9b690c91'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1263.123794] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1263.127391] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1263.127559] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Starting heal instance info cache {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1263.127666] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Rebuilding the list of instances to heal {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1263.366119] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1536f9-f19b-432f-94ad-686da21c9ad1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.373536] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6d047e-0714-4281-943d-074c7f7d9040 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.403101] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b62c1df-dbfa-42c7-9d43-701d644e62e1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.410363] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e042da5-ae8b-4c86-ad1f-fdd210f8b35c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.424246] env[63372]: DEBUG nova.compute.provider_tree [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1263.630822] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Skipping network cache update for instance because it is Building. {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1263.669420] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "refresh_cache-e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1263.669567] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquired lock "refresh_cache-e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1263.669717] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Forcefully refreshing network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1263.669870] env[63372]: DEBUG nova.objects.instance [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lazy-loading 'info_cache' on Instance uuid e922c0c6-50fa-4a4e-9961-bc5bc3a880e7 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1263.820415] env[63372]: DEBUG nova.objects.instance [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lazy-loading 'flavor' on Instance uuid e922c0c6-50fa-4a4e-9961-bc5bc3a880e7 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1263.926792] env[63372]: DEBUG nova.scheduler.client.report [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1264.325374] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e6d62cd7-c472-4b17-95ed-235a253ebb81 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.231s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1264.431535] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.109s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1264.432128] env[63372]: DEBUG nova.compute.manager [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1264.543645] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1264.543901] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1264.936971] env[63372]: DEBUG nova.compute.utils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1264.939063] env[63372]: DEBUG nova.compute.manager [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1264.939063] env[63372]: DEBUG nova.network.neutron [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1264.986899] env[63372]: DEBUG nova.policy [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '228dd399ae7c45ad964fa74c0cbb044c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4b088060a2c4159a62e40e7e627076a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1265.046273] env[63372]: INFO nova.compute.manager [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Detaching volume 0e9d1ac5-508e-4508-8626-deeb9b690c91 [ 1265.076283] env[63372]: INFO nova.virt.block_device [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Attempting to driver detach volume 0e9d1ac5-508e-4508-8626-deeb9b690c91 from mountpoint /dev/sdb [ 1265.076533] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Volume detach. Driver type: vmdk {{(pid=63372) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1265.076722] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227484', 'volume_id': '0e9d1ac5-508e-4508-8626-deeb9b690c91', 'name': 'volume-0e9d1ac5-508e-4508-8626-deeb9b690c91', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e922c0c6-50fa-4a4e-9961-bc5bc3a880e7', 'attached_at': '', 'detached_at': '', 'volume_id': '0e9d1ac5-508e-4508-8626-deeb9b690c91', 'serial': '0e9d1ac5-508e-4508-8626-deeb9b690c91'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1265.077616] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7809138-36a7-43e9-9ced-8d35bd9e0002 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.100565] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9d8883-920d-4474-8d17-38e34e911f3d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.107645] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218b2502-92f9-4656-bc7a-6dafcf7917bf {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.127351] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb5abd9-4363-4f9d-87fc-d2c2905a922e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.141677] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] The volume has not been displaced from its original location: [datastore2] volume-0e9d1ac5-508e-4508-8626-deeb9b690c91/volume-0e9d1ac5-508e-4508-8626-deeb9b690c91.vmdk. No consolidation needed. {{(pid=63372) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1265.146957] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Reconfiguring VM instance instance-00000072 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1265.147249] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-717d3261-00c7-441b-9450-c77bdc27e123 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.164846] env[63372]: DEBUG oslo_vmware.api [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1265.164846] env[63372]: value = "task-1024975" [ 1265.164846] env[63372]: _type = "Task" [ 1265.164846] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.172672] env[63372]: DEBUG oslo_vmware.api [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024975, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.257696] env[63372]: DEBUG nova.network.neutron [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Successfully created port: e4b36a43-ed17-426c-a57f-51dfc8a6ee8d {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1265.390331] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Updating instance_info_cache with network_info: [{"id": "77eef7fd-df07-4cd7-9a4c-0ac51961a494", "address": "fa:16:3e:f3:8f:fc", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77eef7fd-df", "ovs_interfaceid": "77eef7fd-df07-4cd7-9a4c-0ac51961a494", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1265.441584] env[63372]: DEBUG nova.compute.manager [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1265.674622] env[63372]: DEBUG oslo_vmware.api [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024975, 'name': ReconfigVM_Task, 'duration_secs': 0.231923} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.674919] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Reconfigured VM instance instance-00000072 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1265.679493] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65934b82-40af-4ad9-b171-a46d9ac588a5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.694102] env[63372]: DEBUG oslo_vmware.api [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1265.694102] env[63372]: value = "task-1024976" [ 1265.694102] env[63372]: _type = "Task" [ 1265.694102] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.701791] env[63372]: DEBUG oslo_vmware.api [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024976, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.893192] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Releasing lock "refresh_cache-e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1265.893411] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Updated the network info_cache for instance {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1265.893633] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.893834] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.893967] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63372) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1266.127504] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager.update_available_resource {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1266.204775] env[63372]: DEBUG oslo_vmware.api [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024976, 'name': ReconfigVM_Task, 'duration_secs': 0.140513} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.205120] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227484', 'volume_id': '0e9d1ac5-508e-4508-8626-deeb9b690c91', 'name': 'volume-0e9d1ac5-508e-4508-8626-deeb9b690c91', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e922c0c6-50fa-4a4e-9961-bc5bc3a880e7', 'attached_at': '', 'detached_at': '', 'volume_id': '0e9d1ac5-508e-4508-8626-deeb9b690c91', 'serial': '0e9d1ac5-508e-4508-8626-deeb9b690c91'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1266.451780] env[63372]: DEBUG nova.compute.manager [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1266.477070] env[63372]: DEBUG nova.virt.hardware [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1266.477342] env[63372]: DEBUG nova.virt.hardware [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1266.477504] env[63372]: DEBUG nova.virt.hardware [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1266.477692] env[63372]: DEBUG nova.virt.hardware [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1266.477845] env[63372]: DEBUG nova.virt.hardware [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1266.478006] env[63372]: DEBUG nova.virt.hardware [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1266.478224] env[63372]: DEBUG nova.virt.hardware [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1266.478386] env[63372]: DEBUG nova.virt.hardware [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1266.478555] env[63372]: DEBUG nova.virt.hardware [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1266.478719] env[63372]: DEBUG nova.virt.hardware [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1266.478894] env[63372]: DEBUG nova.virt.hardware [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1266.479748] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe639eb8-b194-4153-bfd0-32a4fa049a3e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.489263] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d3b964-594c-4d55-bcb3-1aeeda28ea44 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.631453] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.631707] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.631867] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1266.632039] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63372) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1266.632964] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5a327c-2bd4-42f0-9169-5c8ae55092d9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.638217] env[63372]: DEBUG nova.compute.manager [req-a8dfa6de-efe1-457d-a929-9e52f1c2ddce req-ee62fcff-1246-4428-9acb-4222a5867ab5 service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Received event network-vif-plugged-e4b36a43-ed17-426c-a57f-51dfc8a6ee8d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1266.638451] env[63372]: DEBUG oslo_concurrency.lockutils [req-a8dfa6de-efe1-457d-a929-9e52f1c2ddce req-ee62fcff-1246-4428-9acb-4222a5867ab5 service nova] Acquiring lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.638669] env[63372]: DEBUG oslo_concurrency.lockutils [req-a8dfa6de-efe1-457d-a929-9e52f1c2ddce req-ee62fcff-1246-4428-9acb-4222a5867ab5 service nova] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.638838] env[63372]: DEBUG oslo_concurrency.lockutils [req-a8dfa6de-efe1-457d-a929-9e52f1c2ddce req-ee62fcff-1246-4428-9acb-4222a5867ab5 service nova] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1266.639014] env[63372]: DEBUG nova.compute.manager [req-a8dfa6de-efe1-457d-a929-9e52f1c2ddce req-ee62fcff-1246-4428-9acb-4222a5867ab5 service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] No waiting events found dispatching network-vif-plugged-e4b36a43-ed17-426c-a57f-51dfc8a6ee8d {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1266.639191] env[63372]: WARNING nova.compute.manager [req-a8dfa6de-efe1-457d-a929-9e52f1c2ddce req-ee62fcff-1246-4428-9acb-4222a5867ab5 service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Received unexpected event network-vif-plugged-e4b36a43-ed17-426c-a57f-51dfc8a6ee8d for instance with vm_state building and task_state spawning. [ 1266.645141] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05d1642-139e-4731-8194-21b9af722ff1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.659761] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c71ef179-938a-4042-a0b2-a22e55195b04 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.665977] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb47f69-4bcc-40ad-9ecb-91046c8daf89 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.695827] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181259MB free_disk=186GB free_vcpus=48 pci_devices=None {{(pid=63372) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1266.696029] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.696229] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.720645] env[63372]: DEBUG nova.network.neutron [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Successfully updated port: e4b36a43-ed17-426c-a57f-51dfc8a6ee8d {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1266.747998] env[63372]: DEBUG nova.objects.instance [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lazy-loading 'flavor' on Instance uuid e922c0c6-50fa-4a4e-9961-bc5bc3a880e7 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1267.223438] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "refresh_cache-b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1267.223597] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquired lock "refresh_cache-b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1267.223710] env[63372]: DEBUG nova.network.neutron [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1267.721671] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance e922c0c6-50fa-4a4e-9961-bc5bc3a880e7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1267.721923] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1267.722076] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1267.722274] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1267.754519] env[63372]: DEBUG nova.network.neutron [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1267.756357] env[63372]: DEBUG oslo_concurrency.lockutils [None req-a950c48a-0208-40f9-b2a3-72b65e87568e tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.212s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1267.761333] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-560dba1d-0522-4232-a75e-a4c67c1dbf7c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.768893] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c805c5-3c40-431e-8edb-89d7f7197ebc {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.803152] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d8dd65-cf96-43cf-a754-8485dce2ba95 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.810938] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd255888-50f5-4e1a-8d8e-9af364768746 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.823948] env[63372]: DEBUG nova.compute.provider_tree [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1267.895922] env[63372]: DEBUG nova.network.neutron [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Updating instance_info_cache with network_info: [{"id": "e4b36a43-ed17-426c-a57f-51dfc8a6ee8d", "address": "fa:16:3e:74:73:ce", "network": {"id": "b161f533-292b-431a-a171-e9d61bc90105", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1928954423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4b088060a2c4159a62e40e7e627076a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4b36a43-ed", "ovs_interfaceid": "e4b36a43-ed17-426c-a57f-51dfc8a6ee8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1268.326926] env[63372]: DEBUG nova.scheduler.client.report [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1268.398189] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Releasing lock "refresh_cache-b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1268.398529] env[63372]: DEBUG nova.compute.manager [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Instance network_info: |[{"id": "e4b36a43-ed17-426c-a57f-51dfc8a6ee8d", "address": "fa:16:3e:74:73:ce", "network": {"id": "b161f533-292b-431a-a171-e9d61bc90105", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1928954423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4b088060a2c4159a62e40e7e627076a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4b36a43-ed", "ovs_interfaceid": "e4b36a43-ed17-426c-a57f-51dfc8a6ee8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1268.398976] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:73:ce', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8480e3f5-68bd-4c27-ae1f-7c994a8202b1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e4b36a43-ed17-426c-a57f-51dfc8a6ee8d', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1268.406469] env[63372]: DEBUG oslo.service.loopingcall [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1268.406679] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1268.407104] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1619b7b-0acd-49d6-a833-fb0bbf9b3e68 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.427431] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1268.427431] env[63372]: value = "task-1024977" [ 1268.427431] env[63372]: _type = "Task" [ 1268.427431] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.435622] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024977, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.662520] env[63372]: DEBUG nova.compute.manager [req-359888cc-4e94-4d5f-b6c4-b26276ea5655 req-4489b682-e83e-4b33-ab57-ded4ed8c0e90 service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Received event network-changed-e4b36a43-ed17-426c-a57f-51dfc8a6ee8d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1268.662731] env[63372]: DEBUG nova.compute.manager [req-359888cc-4e94-4d5f-b6c4-b26276ea5655 req-4489b682-e83e-4b33-ab57-ded4ed8c0e90 service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Refreshing instance network info cache due to event network-changed-e4b36a43-ed17-426c-a57f-51dfc8a6ee8d. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1268.662980] env[63372]: DEBUG oslo_concurrency.lockutils [req-359888cc-4e94-4d5f-b6c4-b26276ea5655 req-4489b682-e83e-4b33-ab57-ded4ed8c0e90 service nova] Acquiring lock "refresh_cache-b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1268.663150] env[63372]: DEBUG oslo_concurrency.lockutils [req-359888cc-4e94-4d5f-b6c4-b26276ea5655 req-4489b682-e83e-4b33-ab57-ded4ed8c0e90 service nova] Acquired lock "refresh_cache-b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1268.663317] env[63372]: DEBUG nova.network.neutron [req-359888cc-4e94-4d5f-b6c4-b26276ea5655 req-4489b682-e83e-4b33-ab57-ded4ed8c0e90 service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Refreshing network info cache for port e4b36a43-ed17-426c-a57f-51dfc8a6ee8d {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1268.790017] env[63372]: DEBUG oslo_concurrency.lockutils [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.790355] env[63372]: DEBUG oslo_concurrency.lockutils [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.790457] env[63372]: DEBUG oslo_concurrency.lockutils [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.790638] env[63372]: DEBUG oslo_concurrency.lockutils [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.790808] env[63372]: DEBUG oslo_concurrency.lockutils [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.792998] env[63372]: INFO nova.compute.manager [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Terminating instance [ 1268.794854] env[63372]: DEBUG nova.compute.manager [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1268.795064] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1268.796311] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98eb5f65-f7d9-464e-b9d1-ed6ef404dae7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.803878] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1268.804110] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-97ca106b-cac1-453c-a001-9e6b2461bece {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.810351] env[63372]: DEBUG oslo_vmware.api [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1268.810351] env[63372]: value = "task-1024978" [ 1268.810351] env[63372]: _type = "Task" [ 1268.810351] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.817376] env[63372]: DEBUG oslo_vmware.api [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024978, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.832160] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1268.832347] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.136s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.938215] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024977, 'name': CreateVM_Task, 'duration_secs': 0.308896} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.938441] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1268.945925] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1268.946135] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1268.946485] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1268.946750] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aab73778-85dc-4f95-9283-4ffd734448d3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.952338] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1268.952338] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]524d6006-f3bf-b734-af8c-63b1835ac339" [ 1268.952338] env[63372]: _type = "Task" [ 1268.952338] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.960402] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524d6006-f3bf-b734-af8c-63b1835ac339, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.322361] env[63372]: DEBUG oslo_vmware.api [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024978, 'name': PowerOffVM_Task, 'duration_secs': 0.204733} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.322639] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1269.322841] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1269.323111] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-95c52a94-8747-4cb3-b782-ea06ee8d05e6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.351507] env[63372]: DEBUG nova.network.neutron [req-359888cc-4e94-4d5f-b6c4-b26276ea5655 req-4489b682-e83e-4b33-ab57-ded4ed8c0e90 service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Updated VIF entry in instance network info cache for port e4b36a43-ed17-426c-a57f-51dfc8a6ee8d. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1269.351889] env[63372]: DEBUG nova.network.neutron [req-359888cc-4e94-4d5f-b6c4-b26276ea5655 req-4489b682-e83e-4b33-ab57-ded4ed8c0e90 service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Updating instance_info_cache with network_info: [{"id": "e4b36a43-ed17-426c-a57f-51dfc8a6ee8d", "address": "fa:16:3e:74:73:ce", "network": {"id": "b161f533-292b-431a-a171-e9d61bc90105", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1928954423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4b088060a2c4159a62e40e7e627076a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4b36a43-ed", "ovs_interfaceid": "e4b36a43-ed17-426c-a57f-51dfc8a6ee8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1269.389340] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1269.389525] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Deleting contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1269.389679] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Deleting the datastore file [datastore1] e922c0c6-50fa-4a4e-9961-bc5bc3a880e7 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1269.389942] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-01932075-4d5e-4a60-bbfa-a48252160e72 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.396227] env[63372]: DEBUG oslo_vmware.api [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1269.396227] env[63372]: value = "task-1024980" [ 1269.396227] env[63372]: _type = "Task" [ 1269.396227] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.403662] env[63372]: DEBUG oslo_vmware.api [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024980, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.464571] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]524d6006-f3bf-b734-af8c-63b1835ac339, 'name': SearchDatastore_Task, 'duration_secs': 0.012737} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.464826] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1269.465069] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1269.465308] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1269.465454] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1269.465630] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1269.465880] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8697773-aff3-4319-b80d-b6669c3d00c0 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.481684] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1269.481919] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1269.482750] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a61966c-8887-4c57-bf64-5efb7f6da147 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.487639] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1269.487639] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52b14f82-f8dc-091c-f9c6-b3a8ec259dc2" [ 1269.487639] env[63372]: _type = "Task" [ 1269.487639] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.496492] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b14f82-f8dc-091c-f9c6-b3a8ec259dc2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.855240] env[63372]: DEBUG oslo_concurrency.lockutils [req-359888cc-4e94-4d5f-b6c4-b26276ea5655 req-4489b682-e83e-4b33-ab57-ded4ed8c0e90 service nova] Releasing lock "refresh_cache-b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1269.906346] env[63372]: DEBUG oslo_vmware.api [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024980, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155333} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.906594] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1269.906779] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Deleted contents of the VM from datastore datastore1 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1269.906957] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1269.907144] env[63372]: INFO nova.compute.manager [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1269.907381] env[63372]: DEBUG oslo.service.loopingcall [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1269.907577] env[63372]: DEBUG nova.compute.manager [-] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1269.907668] env[63372]: DEBUG nova.network.neutron [-] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1269.997416] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52b14f82-f8dc-091c-f9c6-b3a8ec259dc2, 'name': SearchDatastore_Task, 'duration_secs': 0.008461} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.998237] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cec266a7-3d81-4303-8cf1-6ec672da0951 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.003461] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1270.003461] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]52c1edf1-db09-91e3-3244-dd65fdc866bd" [ 1270.003461] env[63372]: _type = "Task" [ 1270.003461] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.011626] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c1edf1-db09-91e3-3244-dd65fdc866bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.514934] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]52c1edf1-db09-91e3-3244-dd65fdc866bd, 'name': SearchDatastore_Task, 'duration_secs': 0.010571} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.515298] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1270.515524] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c/b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1270.515777] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2f78a95-174c-4f60-809f-609397b43655 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.524777] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1270.524777] env[63372]: value = "task-1024981" [ 1270.524777] env[63372]: _type = "Task" [ 1270.524777] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.531015] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024981, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.691406] env[63372]: DEBUG nova.compute.manager [req-20598874-2d32-467c-ab2f-5dd9dbb0cec2 req-3ec1a8c7-50d4-4485-b9ba-b5358e1e83e8 service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Received event network-vif-deleted-77eef7fd-df07-4cd7-9a4c-0ac51961a494 {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1270.691406] env[63372]: INFO nova.compute.manager [req-20598874-2d32-467c-ab2f-5dd9dbb0cec2 req-3ec1a8c7-50d4-4485-b9ba-b5358e1e83e8 service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Neutron deleted interface 77eef7fd-df07-4cd7-9a4c-0ac51961a494; detaching it from the instance and deleting it from the info cache [ 1270.691556] env[63372]: DEBUG nova.network.neutron [req-20598874-2d32-467c-ab2f-5dd9dbb0cec2 req-3ec1a8c7-50d4-4485-b9ba-b5358e1e83e8 service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1270.847659] env[63372]: DEBUG nova.network.neutron [-] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1271.032783] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024981, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458255} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.033157] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c/b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1271.033284] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1271.033530] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6e0c14d1-f75f-43ed-9f37-2e581b4f580a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.039309] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1271.039309] env[63372]: value = "task-1024982" [ 1271.039309] env[63372]: _type = "Task" [ 1271.039309] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1271.046749] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024982, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.195520] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a9d75d62-9e05-4b4a-95f6-eb299c432122 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.204650] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cf0d1f-90fa-4780-8b01-f4ee65182162 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.229077] env[63372]: DEBUG nova.compute.manager [req-20598874-2d32-467c-ab2f-5dd9dbb0cec2 req-3ec1a8c7-50d4-4485-b9ba-b5358e1e83e8 service nova] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Detach interface failed, port_id=77eef7fd-df07-4cd7-9a4c-0ac51961a494, reason: Instance e922c0c6-50fa-4a4e-9961-bc5bc3a880e7 could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1271.349993] env[63372]: INFO nova.compute.manager [-] [instance: e922c0c6-50fa-4a4e-9961-bc5bc3a880e7] Took 1.44 seconds to deallocate network for instance. [ 1271.548881] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024982, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070606} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.549166] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1271.549905] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f31d67-744c-4ab5-a84a-ee150d3908de {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.571692] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c/b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1271.571861] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59d2d95c-2e50-4581-b67d-0e69d75cc9a1 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.590497] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1271.590497] env[63372]: value = "task-1024983" [ 1271.590497] env[63372]: _type = "Task" [ 1271.590497] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1271.597923] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024983, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.856823] env[63372]: DEBUG oslo_concurrency.lockutils [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1271.857047] env[63372]: DEBUG oslo_concurrency.lockutils [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1271.857272] env[63372]: DEBUG nova.objects.instance [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lazy-loading 'resources' on Instance uuid e922c0c6-50fa-4a4e-9961-bc5bc3a880e7 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1272.100438] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024983, 'name': ReconfigVM_Task, 'duration_secs': 0.425241} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.100855] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Reconfigured VM instance instance-00000073 to attach disk [datastore2] b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c/b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1272.101392] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6dca9dd7-77de-41e8-8f87-ab93c514e1d9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.107750] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1272.107750] env[63372]: value = "task-1024984" [ 1272.107750] env[63372]: _type = "Task" [ 1272.107750] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.118769] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024984, 'name': Rename_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.399603] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f7d3cc1-69eb-4979-8ac3-48212eaf38ad {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.407206] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab81f106-0a4b-40cd-ba4e-54c78213e263 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.436825] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4867dd0-5d30-4dd0-bca9-b0424b8f1711 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.443504] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3436106c-c5eb-459e-af1f-a575233cfb54 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.456475] env[63372]: DEBUG nova.compute.provider_tree [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1272.616870] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024984, 'name': Rename_Task, 'duration_secs': 0.156292} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.617107] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1272.617345] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d899c0b-e5eb-493f-980a-61be44b5fa12 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.623418] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1272.623418] env[63372]: value = "task-1024985" [ 1272.623418] env[63372]: _type = "Task" [ 1272.623418] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.630401] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024985, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.959604] env[63372]: DEBUG nova.scheduler.client.report [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1273.133995] env[63372]: DEBUG oslo_vmware.api [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024985, 'name': PowerOnVM_Task, 'duration_secs': 0.421076} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1273.134440] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1273.134490] env[63372]: INFO nova.compute.manager [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Took 6.68 seconds to spawn the instance on the hypervisor. [ 1273.134667] env[63372]: DEBUG nova.compute.manager [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1273.135473] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a46656e0-18e4-4c78-8170-f80f99fee82d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.464382] env[63372]: DEBUG oslo_concurrency.lockutils [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.607s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1273.481833] env[63372]: INFO nova.scheduler.client.report [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Deleted allocations for instance e922c0c6-50fa-4a4e-9961-bc5bc3a880e7 [ 1273.651919] env[63372]: INFO nova.compute.manager [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Took 11.34 seconds to build instance. [ 1273.989180] env[63372]: DEBUG oslo_concurrency.lockutils [None req-615e30af-47c9-4e20-be7d-53213ea9e720 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "e922c0c6-50fa-4a4e-9961-bc5bc3a880e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.199s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1274.153645] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b0fa3d1a-eeef-45a6-a555-1fdb81d190a4 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.852s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1274.346613] env[63372]: DEBUG nova.compute.manager [req-6a7e9e96-3c97-4a14-95b6-0ef65921d39c req-8300b298-789b-43e8-a668-1b49fa81954e service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Received event network-changed-e4b36a43-ed17-426c-a57f-51dfc8a6ee8d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1274.346821] env[63372]: DEBUG nova.compute.manager [req-6a7e9e96-3c97-4a14-95b6-0ef65921d39c req-8300b298-789b-43e8-a668-1b49fa81954e service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Refreshing instance network info cache due to event network-changed-e4b36a43-ed17-426c-a57f-51dfc8a6ee8d. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1274.347048] env[63372]: DEBUG oslo_concurrency.lockutils [req-6a7e9e96-3c97-4a14-95b6-0ef65921d39c req-8300b298-789b-43e8-a668-1b49fa81954e service nova] Acquiring lock "refresh_cache-b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1274.347201] env[63372]: DEBUG oslo_concurrency.lockutils [req-6a7e9e96-3c97-4a14-95b6-0ef65921d39c req-8300b298-789b-43e8-a668-1b49fa81954e service nova] Acquired lock "refresh_cache-b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.347368] env[63372]: DEBUG nova.network.neutron [req-6a7e9e96-3c97-4a14-95b6-0ef65921d39c req-8300b298-789b-43e8-a668-1b49fa81954e service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Refreshing network info cache for port e4b36a43-ed17-426c-a57f-51dfc8a6ee8d {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1275.066197] env[63372]: DEBUG nova.network.neutron [req-6a7e9e96-3c97-4a14-95b6-0ef65921d39c req-8300b298-789b-43e8-a668-1b49fa81954e service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Updated VIF entry in instance network info cache for port e4b36a43-ed17-426c-a57f-51dfc8a6ee8d. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1275.066623] env[63372]: DEBUG nova.network.neutron [req-6a7e9e96-3c97-4a14-95b6-0ef65921d39c req-8300b298-789b-43e8-a668-1b49fa81954e service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Updating instance_info_cache with network_info: [{"id": "e4b36a43-ed17-426c-a57f-51dfc8a6ee8d", "address": "fa:16:3e:74:73:ce", "network": {"id": "b161f533-292b-431a-a171-e9d61bc90105", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1928954423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4b088060a2c4159a62e40e7e627076a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4b36a43-ed", "ovs_interfaceid": "e4b36a43-ed17-426c-a57f-51dfc8a6ee8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1275.570596] env[63372]: DEBUG oslo_concurrency.lockutils [req-6a7e9e96-3c97-4a14-95b6-0ef65921d39c req-8300b298-789b-43e8-a668-1b49fa81954e service nova] Releasing lock "refresh_cache-b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1275.808105] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1275.808105] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1276.310097] env[63372]: DEBUG nova.compute.manager [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Starting instance... {{(pid=63372) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1276.829625] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1276.829884] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1276.831441] env[63372]: INFO nova.compute.claims [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1277.942335] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14524168-1e89-47e2-96af-542bc1d1f549 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.949874] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44763792-8c49-4a06-b24f-5bf965caed10 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.979150] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519db6ba-6138-4cde-af6c-3ac04144425c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.985718] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e99e886d-4f5b-42b9-9451-11d3b6714378 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.999472] env[63372]: DEBUG nova.compute.provider_tree [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1278.502726] env[63372]: DEBUG nova.scheduler.client.report [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1279.008047] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.178s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1279.008622] env[63372]: DEBUG nova.compute.manager [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Start building networks asynchronously for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1279.513970] env[63372]: DEBUG nova.compute.utils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1279.515425] env[63372]: DEBUG nova.compute.manager [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Allocating IP information in the background. {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1279.515590] env[63372]: DEBUG nova.network.neutron [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] allocate_for_instance() {{(pid=63372) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1279.563928] env[63372]: DEBUG nova.policy [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc5c59e1419843ad90b1bc33f0a55c31', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8452e71bbbd04222bb6b868017b1b6ed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63372) authorize /opt/stack/nova/nova/policy.py:201}} [ 1279.829106] env[63372]: DEBUG nova.network.neutron [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Successfully created port: b87875d2-f696-402b-8e38-3341cb8effad {{(pid=63372) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1280.019530] env[63372]: DEBUG nova.compute.manager [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Start building block device mappings for instance. {{(pid=63372) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1281.029935] env[63372]: DEBUG nova.compute.manager [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Start spawning the instance on the hypervisor. {{(pid=63372) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1281.056162] env[63372]: DEBUG nova.virt.hardware [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T11:33:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T11:32:43Z,direct_url=,disk_format='vmdk',id=d7ae1717-77b2-47f6-9acd-b27bac221f7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='93ad9fbc5a8646d6b4cf17fa4ff19fe2',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T11:32:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1281.056441] env[63372]: DEBUG nova.virt.hardware [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Flavor limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1281.056603] env[63372]: DEBUG nova.virt.hardware [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Image limits 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1281.056789] env[63372]: DEBUG nova.virt.hardware [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Flavor pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1281.056937] env[63372]: DEBUG nova.virt.hardware [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Image pref 0:0:0 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1281.057105] env[63372]: DEBUG nova.virt.hardware [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63372) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1281.057320] env[63372]: DEBUG nova.virt.hardware [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1281.057478] env[63372]: DEBUG nova.virt.hardware [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1281.057644] env[63372]: DEBUG nova.virt.hardware [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Got 1 possible topologies {{(pid=63372) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1281.057808] env[63372]: DEBUG nova.virt.hardware [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1281.057982] env[63372]: DEBUG nova.virt.hardware [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63372) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1281.058862] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ddc870-45fd-420c-8d6d-58dd34935bab {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.066820] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a229d7-3cd1-4c75-b197-56dee040a463 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.193860] env[63372]: DEBUG nova.compute.manager [req-c665a359-7b56-4433-91e1-d1a12c390bb9 req-ee9b49f3-24b9-4ca4-8139-cff21ce16567 service nova] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Received event network-vif-plugged-b87875d2-f696-402b-8e38-3341cb8effad {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1281.193860] env[63372]: DEBUG oslo_concurrency.lockutils [req-c665a359-7b56-4433-91e1-d1a12c390bb9 req-ee9b49f3-24b9-4ca4-8139-cff21ce16567 service nova] Acquiring lock "3934ef28-318d-40a6-a7fd-becfbdaaa978-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1281.196679] env[63372]: DEBUG oslo_concurrency.lockutils [req-c665a359-7b56-4433-91e1-d1a12c390bb9 req-ee9b49f3-24b9-4ca4-8139-cff21ce16567 service nova] Lock "3934ef28-318d-40a6-a7fd-becfbdaaa978-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1281.196679] env[63372]: DEBUG oslo_concurrency.lockutils [req-c665a359-7b56-4433-91e1-d1a12c390bb9 req-ee9b49f3-24b9-4ca4-8139-cff21ce16567 service nova] Lock "3934ef28-318d-40a6-a7fd-becfbdaaa978-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1281.196679] env[63372]: DEBUG nova.compute.manager [req-c665a359-7b56-4433-91e1-d1a12c390bb9 req-ee9b49f3-24b9-4ca4-8139-cff21ce16567 service nova] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] No waiting events found dispatching network-vif-plugged-b87875d2-f696-402b-8e38-3341cb8effad {{(pid=63372) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1281.196679] env[63372]: WARNING nova.compute.manager [req-c665a359-7b56-4433-91e1-d1a12c390bb9 req-ee9b49f3-24b9-4ca4-8139-cff21ce16567 service nova] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Received unexpected event network-vif-plugged-b87875d2-f696-402b-8e38-3341cb8effad for instance with vm_state building and task_state spawning. [ 1281.277915] env[63372]: DEBUG nova.network.neutron [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Successfully updated port: b87875d2-f696-402b-8e38-3341cb8effad {{(pid=63372) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1281.781555] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "refresh_cache-3934ef28-318d-40a6-a7fd-becfbdaaa978" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1281.781683] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired lock "refresh_cache-3934ef28-318d-40a6-a7fd-becfbdaaa978" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1281.781836] env[63372]: DEBUG nova.network.neutron [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Building network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1282.312371] env[63372]: DEBUG nova.network.neutron [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Instance cache missing network info. {{(pid=63372) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1282.442305] env[63372]: DEBUG nova.network.neutron [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Updating instance_info_cache with network_info: [{"id": "b87875d2-f696-402b-8e38-3341cb8effad", "address": "fa:16:3e:c9:94:d7", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb87875d2-f6", "ovs_interfaceid": "b87875d2-f696-402b-8e38-3341cb8effad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1282.945214] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Releasing lock "refresh_cache-3934ef28-318d-40a6-a7fd-becfbdaaa978" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1282.945548] env[63372]: DEBUG nova.compute.manager [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Instance network_info: |[{"id": "b87875d2-f696-402b-8e38-3341cb8effad", "address": "fa:16:3e:c9:94:d7", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb87875d2-f6", "ovs_interfaceid": "b87875d2-f696-402b-8e38-3341cb8effad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63372) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1282.945994] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:94:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b87875d2-f696-402b-8e38-3341cb8effad', 'vif_model': 'vmxnet3'}] {{(pid=63372) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1282.953564] env[63372]: DEBUG oslo.service.loopingcall [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1282.953791] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Creating VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1282.954034] env[63372]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6d10243-2652-4edc-8d44-ed7e7650f253 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.974224] env[63372]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1282.974224] env[63372]: value = "task-1024986" [ 1282.974224] env[63372]: _type = "Task" [ 1282.974224] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1282.981630] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024986, 'name': CreateVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.220564] env[63372]: DEBUG nova.compute.manager [req-c420b141-feb4-4d48-b8be-363412951d73 req-9162de66-b3de-4117-a71d-31b95760cd7b service nova] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Received event network-changed-b87875d2-f696-402b-8e38-3341cb8effad {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1283.220770] env[63372]: DEBUG nova.compute.manager [req-c420b141-feb4-4d48-b8be-363412951d73 req-9162de66-b3de-4117-a71d-31b95760cd7b service nova] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Refreshing instance network info cache due to event network-changed-b87875d2-f696-402b-8e38-3341cb8effad. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1283.221075] env[63372]: DEBUG oslo_concurrency.lockutils [req-c420b141-feb4-4d48-b8be-363412951d73 req-9162de66-b3de-4117-a71d-31b95760cd7b service nova] Acquiring lock "refresh_cache-3934ef28-318d-40a6-a7fd-becfbdaaa978" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1283.221212] env[63372]: DEBUG oslo_concurrency.lockutils [req-c420b141-feb4-4d48-b8be-363412951d73 req-9162de66-b3de-4117-a71d-31b95760cd7b service nova] Acquired lock "refresh_cache-3934ef28-318d-40a6-a7fd-becfbdaaa978" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.221385] env[63372]: DEBUG nova.network.neutron [req-c420b141-feb4-4d48-b8be-363412951d73 req-9162de66-b3de-4117-a71d-31b95760cd7b service nova] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Refreshing network info cache for port b87875d2-f696-402b-8e38-3341cb8effad {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1283.484549] env[63372]: DEBUG oslo_vmware.api [-] Task: {'id': task-1024986, 'name': CreateVM_Task, 'duration_secs': 0.334825} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.484946] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Created VM on the ESX host {{(pid=63372) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1283.485404] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1283.485575] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.485934] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1283.486246] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad9eb9ac-ee7d-40dd-bd16-97c9c9434f3b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.490392] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1283.490392] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521fbe8f-efdd-f709-ee99-4f5942207991" [ 1283.490392] env[63372]: _type = "Task" [ 1283.490392] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.498208] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521fbe8f-efdd-f709-ee99-4f5942207991, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.901209] env[63372]: DEBUG nova.network.neutron [req-c420b141-feb4-4d48-b8be-363412951d73 req-9162de66-b3de-4117-a71d-31b95760cd7b service nova] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Updated VIF entry in instance network info cache for port b87875d2-f696-402b-8e38-3341cb8effad. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1283.901591] env[63372]: DEBUG nova.network.neutron [req-c420b141-feb4-4d48-b8be-363412951d73 req-9162de66-b3de-4117-a71d-31b95760cd7b service nova] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Updating instance_info_cache with network_info: [{"id": "b87875d2-f696-402b-8e38-3341cb8effad", "address": "fa:16:3e:c9:94:d7", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb87875d2-f6", "ovs_interfaceid": "b87875d2-f696-402b-8e38-3341cb8effad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1284.000756] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521fbe8f-efdd-f709-ee99-4f5942207991, 'name': SearchDatastore_Task, 'duration_secs': 0.009653} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.001048] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1284.001284] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Processing image d7ae1717-77b2-47f6-9acd-b27bac221f7c {{(pid=63372) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1284.001513] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1284.001661] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1284.001839] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1284.002113] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-190bfdfb-aae0-40be-9382-2ed436ce1ef8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.011627] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63372) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1284.011798] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63372) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1284.012534] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3669ebd7-a061-4857-b6fb-dfd1c51e7b54 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.017466] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1284.017466] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]521a3611-be14-b396-8503-40315aad514c" [ 1284.017466] env[63372]: _type = "Task" [ 1284.017466] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.024714] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521a3611-be14-b396-8503-40315aad514c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.403893] env[63372]: DEBUG oslo_concurrency.lockutils [req-c420b141-feb4-4d48-b8be-363412951d73 req-9162de66-b3de-4117-a71d-31b95760cd7b service nova] Releasing lock "refresh_cache-3934ef28-318d-40a6-a7fd-becfbdaaa978" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1284.527719] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]521a3611-be14-b396-8503-40315aad514c, 'name': SearchDatastore_Task, 'duration_secs': 0.008228} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.528512] env[63372]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab646c83-eee3-472f-8dcf-7bffbe9a6861 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.533494] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1284.533494] env[63372]: value = "session[52489336-fd72-e75a-c369-73e4dfedfba5]523a494f-8be8-8cb2-7303-c452bdce79da" [ 1284.533494] env[63372]: _type = "Task" [ 1284.533494] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.540922] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523a494f-8be8-8cb2-7303-c452bdce79da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.044016] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': session[52489336-fd72-e75a-c369-73e4dfedfba5]523a494f-8be8-8cb2-7303-c452bdce79da, 'name': SearchDatastore_Task, 'duration_secs': 0.009268} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.044311] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1285.044575] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 3934ef28-318d-40a6-a7fd-becfbdaaa978/3934ef28-318d-40a6-a7fd-becfbdaaa978.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1285.044846] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21247421-b5ae-4630-85ef-f01c26ff7aa5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.051873] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1285.051873] env[63372]: value = "task-1024987" [ 1285.051873] env[63372]: _type = "Task" [ 1285.051873] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.059540] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024987, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.561483] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024987, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.062974] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024987, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.614151} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.063267] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d7ae1717-77b2-47f6-9acd-b27bac221f7c/d7ae1717-77b2-47f6-9acd-b27bac221f7c.vmdk to [datastore2] 3934ef28-318d-40a6-a7fd-becfbdaaa978/3934ef28-318d-40a6-a7fd-becfbdaaa978.vmdk {{(pid=63372) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1286.063483] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Extending root virtual disk to 1048576 {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1286.063728] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84223c0c-2b8b-4259-bffb-08b2d6e313d9 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.069458] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1286.069458] env[63372]: value = "task-1024988" [ 1286.069458] env[63372]: _type = "Task" [ 1286.069458] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.076739] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024988, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.579234] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024988, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064424} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.579575] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Extended root virtual disk {{(pid=63372) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1286.580301] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38407afb-0732-4ce0-821f-bd4e2389708d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.601598] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Reconfiguring VM instance instance-00000074 to attach disk [datastore2] 3934ef28-318d-40a6-a7fd-becfbdaaa978/3934ef28-318d-40a6-a7fd-becfbdaaa978.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1286.601836] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-848c440a-6533-484a-8a49-8017451f0128 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.619829] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1286.619829] env[63372]: value = "task-1024989" [ 1286.619829] env[63372]: _type = "Task" [ 1286.619829] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.627009] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024989, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.129632] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024989, 'name': ReconfigVM_Task, 'duration_secs': 0.257982} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.129925] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Reconfigured VM instance instance-00000074 to attach disk [datastore2] 3934ef28-318d-40a6-a7fd-becfbdaaa978/3934ef28-318d-40a6-a7fd-becfbdaaa978.vmdk or device None with type sparse {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1287.130584] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-18cac170-bed7-4d24-972d-95b1d0e76582 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.137502] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1287.137502] env[63372]: value = "task-1024990" [ 1287.137502] env[63372]: _type = "Task" [ 1287.137502] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.145180] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024990, 'name': Rename_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.648018] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024990, 'name': Rename_Task, 'duration_secs': 0.143497} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.648367] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Powering on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1287.648636] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c801164-46e7-4024-b649-7c044a1f5b41 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.654642] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1287.654642] env[63372]: value = "task-1024991" [ 1287.654642] env[63372]: _type = "Task" [ 1287.654642] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.661693] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024991, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.165057] env[63372]: DEBUG oslo_vmware.api [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1024991, 'name': PowerOnVM_Task, 'duration_secs': 0.438824} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.165351] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Powered on the VM {{(pid=63372) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1288.165559] env[63372]: INFO nova.compute.manager [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Took 7.14 seconds to spawn the instance on the hypervisor. [ 1288.165744] env[63372]: DEBUG nova.compute.manager [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Checking state {{(pid=63372) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1288.166503] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397320f8-849d-4976-979f-2b129b03b265 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.684219] env[63372]: INFO nova.compute.manager [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Took 11.87 seconds to build instance. [ 1288.848962] env[63372]: DEBUG nova.compute.manager [req-5761ae9d-9e98-4a8b-8283-394594e43089 req-110fa8c6-8068-4af2-a3aa-8c6dc014d7b4 service nova] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Received event network-changed-b87875d2-f696-402b-8e38-3341cb8effad {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1288.848962] env[63372]: DEBUG nova.compute.manager [req-5761ae9d-9e98-4a8b-8283-394594e43089 req-110fa8c6-8068-4af2-a3aa-8c6dc014d7b4 service nova] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Refreshing instance network info cache due to event network-changed-b87875d2-f696-402b-8e38-3341cb8effad. {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1288.849181] env[63372]: DEBUG oslo_concurrency.lockutils [req-5761ae9d-9e98-4a8b-8283-394594e43089 req-110fa8c6-8068-4af2-a3aa-8c6dc014d7b4 service nova] Acquiring lock "refresh_cache-3934ef28-318d-40a6-a7fd-becfbdaaa978" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1288.849328] env[63372]: DEBUG oslo_concurrency.lockutils [req-5761ae9d-9e98-4a8b-8283-394594e43089 req-110fa8c6-8068-4af2-a3aa-8c6dc014d7b4 service nova] Acquired lock "refresh_cache-3934ef28-318d-40a6-a7fd-becfbdaaa978" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1288.849493] env[63372]: DEBUG nova.network.neutron [req-5761ae9d-9e98-4a8b-8283-394594e43089 req-110fa8c6-8068-4af2-a3aa-8c6dc014d7b4 service nova] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Refreshing network info cache for port b87875d2-f696-402b-8e38-3341cb8effad {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1289.186582] env[63372]: DEBUG oslo_concurrency.lockutils [None req-68f77238-e430-4e66-bcad-5f45530bf98b tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.379s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1289.553391] env[63372]: DEBUG nova.network.neutron [req-5761ae9d-9e98-4a8b-8283-394594e43089 req-110fa8c6-8068-4af2-a3aa-8c6dc014d7b4 service nova] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Updated VIF entry in instance network info cache for port b87875d2-f696-402b-8e38-3341cb8effad. {{(pid=63372) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1289.553761] env[63372]: DEBUG nova.network.neutron [req-5761ae9d-9e98-4a8b-8283-394594e43089 req-110fa8c6-8068-4af2-a3aa-8c6dc014d7b4 service nova] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Updating instance_info_cache with network_info: [{"id": "b87875d2-f696-402b-8e38-3341cb8effad", "address": "fa:16:3e:c9:94:d7", "network": {"id": "e99bbf91-ff2e-4513-ba0e-26eb5d15a4ac", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-606884980-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8452e71bbbd04222bb6b868017b1b6ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1dc3a0d3-9578-4c45-bc16-13ed5b84f5b4", "external-id": "nsx-vlan-transportzone-817", "segmentation_id": 817, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb87875d2-f6", "ovs_interfaceid": "b87875d2-f696-402b-8e38-3341cb8effad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1290.056751] env[63372]: DEBUG oslo_concurrency.lockutils [req-5761ae9d-9e98-4a8b-8283-394594e43089 req-110fa8c6-8068-4af2-a3aa-8c6dc014d7b4 service nova] Releasing lock "refresh_cache-3934ef28-318d-40a6-a7fd-becfbdaaa978" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1312.896998] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1312.897434] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1313.401470] env[63372]: DEBUG nova.compute.utils [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1313.905222] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1314.963507] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1314.963902] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1314.964053] env[63372]: INFO nova.compute.manager [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Attaching volume 6b69a92c-c2b4-4c36-b6df-3b6a41dddda4 to /dev/sdb [ 1314.994489] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2ff51b-2a9d-4dbc-99ae-6b7b41beaee6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.001666] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ee931a-2a1c-4344-9442-5c8edcfdcf29 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.014870] env[63372]: DEBUG nova.virt.block_device [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Updating existing volume attachment record: 0fbb41da-3a98-499b-af62-b611e1903666 {{(pid=63372) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1319.557680] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Volume attach. Driver type: vmdk {{(pid=63372) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1319.557948] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227487', 'volume_id': '6b69a92c-c2b4-4c36-b6df-3b6a41dddda4', 'name': 'volume-6b69a92c-c2b4-4c36-b6df-3b6a41dddda4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c', 'attached_at': '', 'detached_at': '', 'volume_id': '6b69a92c-c2b4-4c36-b6df-3b6a41dddda4', 'serial': '6b69a92c-c2b4-4c36-b6df-3b6a41dddda4'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1319.558822] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b9ad0d-673f-40bb-acc7-c6c329a281ea {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.575395] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34b4147-9193-49bc-a4f6-2b6ba7de405d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.600320] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] volume-6b69a92c-c2b4-4c36-b6df-3b6a41dddda4/volume-6b69a92c-c2b4-4c36-b6df-3b6a41dddda4.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1319.600554] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2745f1c-83bf-4162-9c67-f80ecfd01e17 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.617803] env[63372]: DEBUG oslo_vmware.api [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1319.617803] env[63372]: value = "task-1024994" [ 1319.617803] env[63372]: _type = "Task" [ 1319.617803] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.629279] env[63372]: DEBUG oslo_vmware.api [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024994, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.128364] env[63372]: DEBUG oslo_vmware.api [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024994, 'name': ReconfigVM_Task, 'duration_secs': 0.319118} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.128691] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Reconfigured VM instance instance-00000073 to attach disk [datastore2] volume-6b69a92c-c2b4-4c36-b6df-3b6a41dddda4/volume-6b69a92c-c2b4-4c36-b6df-3b6a41dddda4.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1320.133378] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a004e7a8-d8a3-4535-add3-ea5f088c90a6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.147559] env[63372]: DEBUG oslo_vmware.api [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1320.147559] env[63372]: value = "task-1024995" [ 1320.147559] env[63372]: _type = "Task" [ 1320.147559] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.155071] env[63372]: DEBUG oslo_vmware.api [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024995, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.657640] env[63372]: DEBUG oslo_vmware.api [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024995, 'name': ReconfigVM_Task, 'duration_secs': 0.147794} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.657948] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227487', 'volume_id': '6b69a92c-c2b4-4c36-b6df-3b6a41dddda4', 'name': 'volume-6b69a92c-c2b4-4c36-b6df-3b6a41dddda4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c', 'attached_at': '', 'detached_at': '', 'volume_id': '6b69a92c-c2b4-4c36-b6df-3b6a41dddda4', 'serial': '6b69a92c-c2b4-4c36-b6df-3b6a41dddda4'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1321.695943] env[63372]: DEBUG nova.objects.instance [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lazy-loading 'flavor' on Instance uuid b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1322.204410] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e2120052-35cc-4396-8dd4-868ab32206d0 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.240s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1322.987865] env[63372]: DEBUG oslo_concurrency.lockutils [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1322.988128] env[63372]: DEBUG oslo_concurrency.lockutils [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1323.492035] env[63372]: DEBUG nova.compute.utils [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1323.832959] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1323.833162] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1323.833304] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Starting heal instance info cache {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1323.833463] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Rebuilding the list of instances to heal {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1323.995827] env[63372]: DEBUG oslo_concurrency.lockutils [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1324.363433] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "refresh_cache-b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1324.363610] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquired lock "refresh_cache-b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1324.363762] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Forcefully refreshing network info cache for instance {{(pid=63372) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1324.363919] env[63372]: DEBUG nova.objects.instance [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lazy-loading 'info_cache' on Instance uuid b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1325.050632] env[63372]: DEBUG oslo_concurrency.lockutils [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1325.051033] env[63372]: DEBUG oslo_concurrency.lockutils [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1325.051209] env[63372]: INFO nova.compute.manager [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Attaching volume 4abfccda-7b9a-46bb-97b9-f4e1fb70275a to /dev/sdc [ 1325.081405] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58fab81e-e187-4289-8ea4-f32e15b8f7b8 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.088820] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a15db2-adec-4f9c-b003-43fdf3782ae5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.101484] env[63372]: DEBUG nova.virt.block_device [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Updating existing volume attachment record: 9b81b5be-f138-4f9e-b6ad-8ab20fe33728 {{(pid=63372) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1326.083695] env[63372]: DEBUG nova.network.neutron [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Updating instance_info_cache with network_info: [{"id": "e4b36a43-ed17-426c-a57f-51dfc8a6ee8d", "address": "fa:16:3e:74:73:ce", "network": {"id": "b161f533-292b-431a-a171-e9d61bc90105", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1928954423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4b088060a2c4159a62e40e7e627076a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4b36a43-ed", "ovs_interfaceid": "e4b36a43-ed17-426c-a57f-51dfc8a6ee8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1326.586521] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Releasing lock "refresh_cache-b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" {{(pid=63372) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1326.586733] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Updated the network info_cache for instance {{(pid=63372) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1326.586973] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.587153] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.587300] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.587461] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.587607] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.587750] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.587947] env[63372]: DEBUG nova.compute.manager [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63372) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1326.814327] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1326.814578] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1327.127651] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager.update_available_resource {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1327.318012] env[63372]: DEBUG nova.compute.utils [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Using /dev/sd instead of None {{(pid=63372) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1327.630211] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1327.630448] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1327.630624] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1327.630781] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63372) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1327.631753] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50c0020-4aff-478d-8f14-0983154f8e41 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.639584] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383179b4-009e-44da-9e40-32c66966af0e {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.654140] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34663c45-4153-4e91-905b-8af4b2a738ff {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.660313] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57f1ebc-9462-42aa-8418-722dc0fdbcf7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.688462] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181208MB free_disk=186GB free_vcpus=48 pci_devices=None {{(pid=63372) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1327.688599] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1327.688786] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1327.821247] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1328.714830] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1328.715103] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Instance 3934ef28-318d-40a6-a7fd-becfbdaaa978 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63372) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1328.715271] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1328.715364] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=63372) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1328.748909] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc24592-5790-4cdd-84d7-e373c6c9c21a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.756670] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6fcc23-622a-4d1a-bd09-5177fed6e5da {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.786959] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d1dc23-b594-43d3-8e0d-0ec712181d0a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.793498] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a75575-daec-4150-985f-292947d40f1f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.805970] env[63372]: DEBUG nova.compute.provider_tree [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1328.874239] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1328.874511] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1328.874755] env[63372]: INFO nova.compute.manager [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Attaching volume 9be502f0-9bc9-4ad1-abc3-2f053bfaf472 to /dev/sdb [ 1328.903778] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6edb1c-553c-4ee9-aff0-f8ba5cb59406 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.910388] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae590d47-7c7b-4fdf-9e0f-ff4baac7b920 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.923074] env[63372]: DEBUG nova.virt.block_device [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Updating existing volume attachment record: fd428e3a-bd6f-4590-b275-cf9fac5009d0 {{(pid=63372) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1329.309473] env[63372]: DEBUG nova.scheduler.client.report [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1329.644573] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Volume attach. Driver type: vmdk {{(pid=63372) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1329.644837] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227488', 'volume_id': '4abfccda-7b9a-46bb-97b9-f4e1fb70275a', 'name': 'volume-4abfccda-7b9a-46bb-97b9-f4e1fb70275a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c', 'attached_at': '', 'detached_at': '', 'volume_id': '4abfccda-7b9a-46bb-97b9-f4e1fb70275a', 'serial': '4abfccda-7b9a-46bb-97b9-f4e1fb70275a'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1329.645715] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e120b3c-9cb8-403f-9227-a3b7f52b1b07 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.661594] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c6944cb-11e6-4215-8fb3-4e92114ddf2f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.687542] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] volume-4abfccda-7b9a-46bb-97b9-f4e1fb70275a/volume-4abfccda-7b9a-46bb-97b9-f4e1fb70275a.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1329.687799] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8354283c-4847-4e8a-b069-51dbd18b490d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.705128] env[63372]: DEBUG oslo_vmware.api [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1329.705128] env[63372]: value = "task-1024999" [ 1329.705128] env[63372]: _type = "Task" [ 1329.705128] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.712573] env[63372]: DEBUG oslo_vmware.api [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024999, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.814705] env[63372]: DEBUG nova.compute.resource_tracker [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63372) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1329.815109] env[63372]: DEBUG oslo_concurrency.lockutils [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.126s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1330.217221] env[63372]: DEBUG oslo_vmware.api [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1024999, 'name': ReconfigVM_Task, 'duration_secs': 0.327589} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.217537] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Reconfigured VM instance instance-00000073 to attach disk [datastore2] volume-4abfccda-7b9a-46bb-97b9-f4e1fb70275a/volume-4abfccda-7b9a-46bb-97b9-f4e1fb70275a.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1330.222095] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6da3e419-6bc0-4890-bc3c-4665e3f5acc7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.236792] env[63372]: DEBUG oslo_vmware.api [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1330.236792] env[63372]: value = "task-1025000" [ 1330.236792] env[63372]: _type = "Task" [ 1330.236792] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.244283] env[63372]: DEBUG oslo_vmware.api [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1025000, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.747254] env[63372]: DEBUG oslo_vmware.api [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1025000, 'name': ReconfigVM_Task, 'duration_secs': 0.149547} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.747599] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227488', 'volume_id': '4abfccda-7b9a-46bb-97b9-f4e1fb70275a', 'name': 'volume-4abfccda-7b9a-46bb-97b9-f4e1fb70275a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c', 'attached_at': '', 'detached_at': '', 'volume_id': '4abfccda-7b9a-46bb-97b9-f4e1fb70275a', 'serial': '4abfccda-7b9a-46bb-97b9-f4e1fb70275a'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1331.784225] env[63372]: DEBUG nova.objects.instance [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lazy-loading 'flavor' on Instance uuid b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1332.290546] env[63372]: DEBUG oslo_concurrency.lockutils [None req-016181c2-faf8-45b9-93f5-0f0bf62d46a1 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.240s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1332.550197] env[63372]: DEBUG oslo_concurrency.lockutils [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1332.550440] env[63372]: DEBUG oslo_concurrency.lockutils [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1333.053799] env[63372]: INFO nova.compute.manager [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Detaching volume 6b69a92c-c2b4-4c36-b6df-3b6a41dddda4 [ 1333.082367] env[63372]: INFO nova.virt.block_device [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Attempting to driver detach volume 6b69a92c-c2b4-4c36-b6df-3b6a41dddda4 from mountpoint /dev/sdb [ 1333.082638] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Volume detach. Driver type: vmdk {{(pid=63372) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1333.082822] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227487', 'volume_id': '6b69a92c-c2b4-4c36-b6df-3b6a41dddda4', 'name': 'volume-6b69a92c-c2b4-4c36-b6df-3b6a41dddda4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c', 'attached_at': '', 'detached_at': '', 'volume_id': '6b69a92c-c2b4-4c36-b6df-3b6a41dddda4', 'serial': '6b69a92c-c2b4-4c36-b6df-3b6a41dddda4'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1333.083728] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aedbd2fe-5496-4590-bb0a-d1223172acf4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.106968] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb85be5-fa70-4d43-820f-99e678c264ee {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.114245] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22a8e3f9-0067-4d90-8411-71d89ec91c01 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.135861] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd90b0f-fd97-4243-a700-84763161e572 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.149517] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] The volume has not been displaced from its original location: [datastore2] volume-6b69a92c-c2b4-4c36-b6df-3b6a41dddda4/volume-6b69a92c-c2b4-4c36-b6df-3b6a41dddda4.vmdk. No consolidation needed. {{(pid=63372) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1333.154751] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Reconfiguring VM instance instance-00000073 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1333.155015] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f465356a-fdda-4b75-950f-be6053328146 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.171872] env[63372]: DEBUG oslo_vmware.api [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1333.171872] env[63372]: value = "task-1025002" [ 1333.171872] env[63372]: _type = "Task" [ 1333.171872] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.179054] env[63372]: DEBUG oslo_vmware.api [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1025002, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.463964] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Volume attach. Driver type: vmdk {{(pid=63372) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1333.464236] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227489', 'volume_id': '9be502f0-9bc9-4ad1-abc3-2f053bfaf472', 'name': 'volume-9be502f0-9bc9-4ad1-abc3-2f053bfaf472', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3934ef28-318d-40a6-a7fd-becfbdaaa978', 'attached_at': '', 'detached_at': '', 'volume_id': '9be502f0-9bc9-4ad1-abc3-2f053bfaf472', 'serial': '9be502f0-9bc9-4ad1-abc3-2f053bfaf472'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1333.465162] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c494d8a0-ef31-414c-8e5a-b2bbaf9f5f3c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.480816] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff45282a-b07d-4eed-b5f8-540515c8049a {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.504104] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] volume-9be502f0-9bc9-4ad1-abc3-2f053bfaf472/volume-9be502f0-9bc9-4ad1-abc3-2f053bfaf472.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1333.504332] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7cb3763-e186-4531-98ad-105c5a7660fd {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.521829] env[63372]: DEBUG oslo_vmware.api [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1333.521829] env[63372]: value = "task-1025003" [ 1333.521829] env[63372]: _type = "Task" [ 1333.521829] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.528993] env[63372]: DEBUG oslo_vmware.api [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1025003, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.681795] env[63372]: DEBUG oslo_vmware.api [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1025002, 'name': ReconfigVM_Task, 'duration_secs': 0.20308} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1333.682106] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Reconfigured VM instance instance-00000073 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1333.686908] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73e47233-23d0-42bc-ace3-83bf386ae645 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.701984] env[63372]: DEBUG oslo_vmware.api [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1333.701984] env[63372]: value = "task-1025004" [ 1333.701984] env[63372]: _type = "Task" [ 1333.701984] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.710989] env[63372]: DEBUG oslo_vmware.api [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1025004, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.031801] env[63372]: DEBUG oslo_vmware.api [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1025003, 'name': ReconfigVM_Task, 'duration_secs': 0.327207} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.032059] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Reconfigured VM instance instance-00000074 to attach disk [datastore1] volume-9be502f0-9bc9-4ad1-abc3-2f053bfaf472/volume-9be502f0-9bc9-4ad1-abc3-2f053bfaf472.vmdk or device None with type thin {{(pid=63372) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1334.036780] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71cbe186-2ad4-4af4-9e6a-e9c7d68664a3 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.051047] env[63372]: DEBUG oslo_vmware.api [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1334.051047] env[63372]: value = "task-1025005" [ 1334.051047] env[63372]: _type = "Task" [ 1334.051047] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.058362] env[63372]: DEBUG oslo_vmware.api [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1025005, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.212016] env[63372]: DEBUG oslo_vmware.api [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1025004, 'name': ReconfigVM_Task, 'duration_secs': 0.15401} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.212343] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227487', 'volume_id': '6b69a92c-c2b4-4c36-b6df-3b6a41dddda4', 'name': 'volume-6b69a92c-c2b4-4c36-b6df-3b6a41dddda4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c', 'attached_at': '', 'detached_at': '', 'volume_id': '6b69a92c-c2b4-4c36-b6df-3b6a41dddda4', 'serial': '6b69a92c-c2b4-4c36-b6df-3b6a41dddda4'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1334.560365] env[63372]: DEBUG oslo_vmware.api [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1025005, 'name': ReconfigVM_Task, 'duration_secs': 0.142149} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.560668] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227489', 'volume_id': '9be502f0-9bc9-4ad1-abc3-2f053bfaf472', 'name': 'volume-9be502f0-9bc9-4ad1-abc3-2f053bfaf472', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3934ef28-318d-40a6-a7fd-becfbdaaa978', 'attached_at': '', 'detached_at': '', 'volume_id': '9be502f0-9bc9-4ad1-abc3-2f053bfaf472', 'serial': '9be502f0-9bc9-4ad1-abc3-2f053bfaf472'} {{(pid=63372) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1334.751408] env[63372]: DEBUG nova.objects.instance [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lazy-loading 'flavor' on Instance uuid b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1335.595550] env[63372]: DEBUG nova.objects.instance [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lazy-loading 'flavor' on Instance uuid 3934ef28-318d-40a6-a7fd-becfbdaaa978 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1335.759172] env[63372]: DEBUG oslo_concurrency.lockutils [None req-92ffa9b5-cd28-44e6-b670-688b22837906 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.209s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1335.784027] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1335.784297] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1336.100126] env[63372]: DEBUG oslo_concurrency.lockutils [None req-9f832f49-2678-4db5-8da2-7ce2471000d4 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.225s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1336.287378] env[63372]: INFO nova.compute.manager [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Detaching volume 4abfccda-7b9a-46bb-97b9-f4e1fb70275a [ 1336.316465] env[63372]: INFO nova.virt.block_device [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Attempting to driver detach volume 4abfccda-7b9a-46bb-97b9-f4e1fb70275a from mountpoint /dev/sdc [ 1336.316723] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Volume detach. Driver type: vmdk {{(pid=63372) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1336.316912] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227488', 'volume_id': '4abfccda-7b9a-46bb-97b9-f4e1fb70275a', 'name': 'volume-4abfccda-7b9a-46bb-97b9-f4e1fb70275a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c', 'attached_at': '', 'detached_at': '', 'volume_id': '4abfccda-7b9a-46bb-97b9-f4e1fb70275a', 'serial': '4abfccda-7b9a-46bb-97b9-f4e1fb70275a'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1336.317771] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23d2d9e4-4a3e-4b3e-928b-dfc572f1516d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.339791] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8261ded0-20a4-41f6-8793-98a42c103c96 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.346637] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee1401d-a890-4a0f-ae49-b69dcddb6347 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.366872] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d76ff89-ae19-4839-88a0-1a61ea12c79f {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.381103] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] The volume has not been displaced from its original location: [datastore2] volume-4abfccda-7b9a-46bb-97b9-f4e1fb70275a/volume-4abfccda-7b9a-46bb-97b9-f4e1fb70275a.vmdk. No consolidation needed. {{(pid=63372) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1336.386309] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Reconfiguring VM instance instance-00000073 to detach disk 2002 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1336.386561] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a69274f3-a313-4c53-ab88-63e3f857fde2 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.404152] env[63372]: DEBUG oslo_vmware.api [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1336.404152] env[63372]: value = "task-1025006" [ 1336.404152] env[63372]: _type = "Task" [ 1336.404152] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.411091] env[63372]: DEBUG oslo_vmware.api [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1025006, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.525894] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1336.526187] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1336.810645] env[63372]: DEBUG oslo_service.periodic_task [None req-b69fa1d1-4058-4a4f-8707-4ea43a28818a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63372) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1336.914069] env[63372]: DEBUG oslo_vmware.api [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1025006, 'name': ReconfigVM_Task, 'duration_secs': 0.225343} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1336.914355] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Reconfigured VM instance instance-00000073 to detach disk 2002 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1336.918943] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bdea3892-0914-46dc-9dee-5fefc0f0e61c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.934866] env[63372]: DEBUG oslo_vmware.api [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1336.934866] env[63372]: value = "task-1025007" [ 1336.934866] env[63372]: _type = "Task" [ 1336.934866] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.942452] env[63372]: DEBUG oslo_vmware.api [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1025007, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.028917] env[63372]: INFO nova.compute.manager [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Detaching volume 9be502f0-9bc9-4ad1-abc3-2f053bfaf472 [ 1337.060009] env[63372]: INFO nova.virt.block_device [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Attempting to driver detach volume 9be502f0-9bc9-4ad1-abc3-2f053bfaf472 from mountpoint /dev/sdb [ 1337.060302] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Volume detach. Driver type: vmdk {{(pid=63372) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1337.060508] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227489', 'volume_id': '9be502f0-9bc9-4ad1-abc3-2f053bfaf472', 'name': 'volume-9be502f0-9bc9-4ad1-abc3-2f053bfaf472', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3934ef28-318d-40a6-a7fd-becfbdaaa978', 'attached_at': '', 'detached_at': '', 'volume_id': '9be502f0-9bc9-4ad1-abc3-2f053bfaf472', 'serial': '9be502f0-9bc9-4ad1-abc3-2f053bfaf472'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1337.061403] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88f7d7e-b621-4c00-8e1c-a5b109e12f00 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.084062] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a67a51b-6776-45c7-9456-023b2799b434 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.090433] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f7eb54-7f5b-4b20-a364-023a58a8afde {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.109496] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f5118fa-bf76-4111-a67b-1a1ac57746a4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.123391] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] The volume has not been displaced from its original location: [datastore1] volume-9be502f0-9bc9-4ad1-abc3-2f053bfaf472/volume-9be502f0-9bc9-4ad1-abc3-2f053bfaf472.vmdk. No consolidation needed. {{(pid=63372) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1337.128599] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Reconfiguring VM instance instance-00000074 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1337.128851] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7196d664-961f-4d29-b3f8-ebb6c8d5e576 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.146321] env[63372]: DEBUG oslo_vmware.api [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1337.146321] env[63372]: value = "task-1025008" [ 1337.146321] env[63372]: _type = "Task" [ 1337.146321] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.153383] env[63372]: DEBUG oslo_vmware.api [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1025008, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.444196] env[63372]: DEBUG oslo_vmware.api [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1025007, 'name': ReconfigVM_Task, 'duration_secs': 0.133307} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.444491] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227488', 'volume_id': '4abfccda-7b9a-46bb-97b9-f4e1fb70275a', 'name': 'volume-4abfccda-7b9a-46bb-97b9-f4e1fb70275a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c', 'attached_at': '', 'detached_at': '', 'volume_id': '4abfccda-7b9a-46bb-97b9-f4e1fb70275a', 'serial': '4abfccda-7b9a-46bb-97b9-f4e1fb70275a'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1337.656224] env[63372]: DEBUG oslo_vmware.api [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1025008, 'name': ReconfigVM_Task, 'duration_secs': 0.192185} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.656496] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Reconfigured VM instance instance-00000074 to detach disk 2001 {{(pid=63372) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1337.661066] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e60aad6e-1499-4b40-93a5-54a9cc57c65b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.675858] env[63372]: DEBUG oslo_vmware.api [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1337.675858] env[63372]: value = "task-1025009" [ 1337.675858] env[63372]: _type = "Task" [ 1337.675858] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.683049] env[63372]: DEBUG oslo_vmware.api [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1025009, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.984409] env[63372]: DEBUG nova.objects.instance [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lazy-loading 'flavor' on Instance uuid b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1338.185756] env[63372]: DEBUG oslo_vmware.api [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1025009, 'name': ReconfigVM_Task, 'duration_secs': 0.132853} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.186089] env[63372]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-227489', 'volume_id': '9be502f0-9bc9-4ad1-abc3-2f053bfaf472', 'name': 'volume-9be502f0-9bc9-4ad1-abc3-2f053bfaf472', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3934ef28-318d-40a6-a7fd-becfbdaaa978', 'attached_at': '', 'detached_at': '', 'volume_id': '9be502f0-9bc9-4ad1-abc3-2f053bfaf472', 'serial': '9be502f0-9bc9-4ad1-abc3-2f053bfaf472'} {{(pid=63372) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1338.725112] env[63372]: DEBUG nova.objects.instance [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lazy-loading 'flavor' on Instance uuid 3934ef28-318d-40a6-a7fd-becfbdaaa978 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1338.992181] env[63372]: DEBUG oslo_concurrency.lockutils [None req-d0a54a9d-15fc-4f72-9fa9-2efedd52ec97 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.208s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1339.731697] env[63372]: DEBUG oslo_concurrency.lockutils [None req-dd543e11-60c6-4c2c-9ae5-5bb314a335db tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.205s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1340.136932] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1340.137342] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1340.137413] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1340.137620] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1340.137796] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1340.140041] env[63372]: INFO nova.compute.manager [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Terminating instance [ 1340.141879] env[63372]: DEBUG nova.compute.manager [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1340.142110] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1340.142940] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1902589a-e68f-4562-9aac-a711bdbc87a4 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.150681] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1340.150907] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fef54442-3214-4066-9870-f1a2cd25e7f6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.157154] env[63372]: DEBUG oslo_vmware.api [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1340.157154] env[63372]: value = "task-1025010" [ 1340.157154] env[63372]: _type = "Task" [ 1340.157154] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.164790] env[63372]: DEBUG oslo_vmware.api [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1025010, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.667259] env[63372]: DEBUG oslo_vmware.api [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1025010, 'name': PowerOffVM_Task, 'duration_secs': 0.187918} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.667526] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1340.667712] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1340.667957] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d8842db8-35f6-4002-9b95-9a923d5804f6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.738067] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1340.738324] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1340.738516] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Deleting the datastore file [datastore2] b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1340.738787] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dfd6cd99-355e-42c4-a53d-2846a435fa4b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.744785] env[63372]: DEBUG oslo_vmware.api [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for the task: (returnval){ [ 1340.744785] env[63372]: value = "task-1025012" [ 1340.744785] env[63372]: _type = "Task" [ 1340.744785] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.752258] env[63372]: DEBUG oslo_vmware.api [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1025012, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.759706] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1340.759929] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1340.760146] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "3934ef28-318d-40a6-a7fd-becfbdaaa978-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1340.760334] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "3934ef28-318d-40a6-a7fd-becfbdaaa978-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1340.760503] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "3934ef28-318d-40a6-a7fd-becfbdaaa978-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1340.762624] env[63372]: INFO nova.compute.manager [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Terminating instance [ 1340.764481] env[63372]: DEBUG nova.compute.manager [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Start destroying the instance on the hypervisor. {{(pid=63372) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1340.764698] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Destroying instance {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1340.765435] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a5a721-e0e5-40f3-b875-c4f9b6b2cfbb {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.771707] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Powering off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1340.771924] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-89fe5724-2338-4291-b9cf-082b6e83722d {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.777384] env[63372]: DEBUG oslo_vmware.api [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1340.777384] env[63372]: value = "task-1025013" [ 1340.777384] env[63372]: _type = "Task" [ 1340.777384] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.785284] env[63372]: DEBUG oslo_vmware.api [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1025013, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.254780] env[63372]: DEBUG oslo_vmware.api [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Task: {'id': task-1025012, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13349} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.255304] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1341.255304] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1341.255435] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1341.255602] env[63372]: INFO nova.compute.manager [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1341.255943] env[63372]: DEBUG oslo.service.loopingcall [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1341.256194] env[63372]: DEBUG nova.compute.manager [-] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1341.256301] env[63372]: DEBUG nova.network.neutron [-] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1341.292630] env[63372]: DEBUG oslo_vmware.api [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1025013, 'name': PowerOffVM_Task, 'duration_secs': 0.174306} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.293032] env[63372]: DEBUG nova.virt.vmwareapi.vm_util [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Powered off the VM {{(pid=63372) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1341.293345] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Unregistering the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1341.293729] env[63372]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4d692059-36e2-496a-b8ca-a1a65a118289 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.361795] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Unregistered the VM {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1341.362044] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Deleting contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1341.362236] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Deleting the datastore file [datastore2] 3934ef28-318d-40a6-a7fd-becfbdaaa978 {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1341.362500] env[63372]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7ae3423-0e2b-4f4d-8669-343e6326b5ae {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.368957] env[63372]: DEBUG oslo_vmware.api [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for the task: (returnval){ [ 1341.368957] env[63372]: value = "task-1025015" [ 1341.368957] env[63372]: _type = "Task" [ 1341.368957] env[63372]: } to complete. {{(pid=63372) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.376552] env[63372]: DEBUG oslo_vmware.api [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1025015, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.710037] env[63372]: DEBUG nova.compute.manager [req-95209c33-a423-4193-989f-41c7c7036995 req-3c1af91e-f900-4375-bd14-5c75641c327a service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Received event network-vif-deleted-e4b36a43-ed17-426c-a57f-51dfc8a6ee8d {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1341.710168] env[63372]: INFO nova.compute.manager [req-95209c33-a423-4193-989f-41c7c7036995 req-3c1af91e-f900-4375-bd14-5c75641c327a service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Neutron deleted interface e4b36a43-ed17-426c-a57f-51dfc8a6ee8d; detaching it from the instance and deleting it from the info cache [ 1341.710337] env[63372]: DEBUG nova.network.neutron [req-95209c33-a423-4193-989f-41c7c7036995 req-3c1af91e-f900-4375-bd14-5c75641c327a service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1341.878919] env[63372]: DEBUG oslo_vmware.api [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Task: {'id': task-1025015, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132419} completed successfully. {{(pid=63372) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.879231] env[63372]: DEBUG nova.virt.vmwareapi.ds_util [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Deleted the datastore file {{(pid=63372) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1341.879424] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Deleted contents of the VM from datastore datastore2 {{(pid=63372) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1341.879612] env[63372]: DEBUG nova.virt.vmwareapi.vmops [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Instance destroyed {{(pid=63372) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1341.879787] env[63372]: INFO nova.compute.manager [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1341.880038] env[63372]: DEBUG oslo.service.loopingcall [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63372) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1341.880239] env[63372]: DEBUG nova.compute.manager [-] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Deallocating network for instance {{(pid=63372) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1341.880332] env[63372]: DEBUG nova.network.neutron [-] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] deallocate_for_instance() {{(pid=63372) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1342.188014] env[63372]: DEBUG nova.network.neutron [-] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1342.212500] env[63372]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8caeb910-51e7-4125-948c-ab0fe4dd5705 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.222363] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b787d03-d58f-4849-aa4a-1a4c65e0cd11 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.247070] env[63372]: DEBUG nova.compute.manager [req-95209c33-a423-4193-989f-41c7c7036995 req-3c1af91e-f900-4375-bd14-5c75641c327a service nova] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Detach interface failed, port_id=e4b36a43-ed17-426c-a57f-51dfc8a6ee8d, reason: Instance b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c could not be found. {{(pid=63372) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1342.691587] env[63372]: INFO nova.compute.manager [-] [instance: b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c] Took 1.44 seconds to deallocate network for instance. [ 1342.768471] env[63372]: DEBUG nova.network.neutron [-] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Updating instance_info_cache with network_info: [] {{(pid=63372) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1343.198814] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1343.199110] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1343.199338] env[63372]: DEBUG nova.objects.instance [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lazy-loading 'resources' on Instance uuid b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1343.270569] env[63372]: INFO nova.compute.manager [-] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Took 1.39 seconds to deallocate network for instance. [ 1343.737164] env[63372]: DEBUG nova.compute.manager [req-a2ebb1f3-fd29-4b78-b5a5-7372f9919cf5 req-2c5bceb3-bd09-49e5-b6a4-b239d36c629f service nova] [instance: 3934ef28-318d-40a6-a7fd-becfbdaaa978] Received event network-vif-deleted-b87875d2-f696-402b-8e38-3341cb8effad {{(pid=63372) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1343.741305] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3109cdc-4a2b-473f-971e-f7dedf6e6137 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.749533] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4076cd1f-dc43-4a9d-95fe-e261c20aa970 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.778563] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1343.779385] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a35a70-019b-4731-8b6a-79885782594c {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.785939] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9467e801-c641-41f3-afee-e5e50fd95fa5 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.798511] env[63372]: DEBUG nova.compute.provider_tree [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1344.301946] env[63372]: DEBUG nova.scheduler.client.report [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1344.806540] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.607s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1344.809795] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.031s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1344.810130] env[63372]: DEBUG nova.objects.instance [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lazy-loading 'resources' on Instance uuid 3934ef28-318d-40a6-a7fd-becfbdaaa978 {{(pid=63372) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1344.836145] env[63372]: INFO nova.scheduler.client.report [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Deleted allocations for instance b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c [ 1345.339490] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce55398-649f-4483-8c0e-767bfe227740 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.347439] env[63372]: DEBUG oslo_concurrency.lockutils [None req-cf9874f8-1531-477b-a1bb-668c0315ab30 tempest-AttachVolumeTestJSON-638311630 tempest-AttachVolumeTestJSON-638311630-project-member] Lock "b5ff3b19-e1a5-42c3-b2c0-ee5e4e254e3c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.210s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1345.353197] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0421fd5b-5a27-44b7-be09-d813c746bed6 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.407486] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f607b875-3cc9-4875-af6f-b62e6001769b {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.418335] env[63372]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ac337d-3f39-4bad-a433-6c1df6a077e7 {{(pid=63372) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.432767] env[63372]: DEBUG nova.compute.provider_tree [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Inventory has not changed in ProviderTree for provider: c1b110ca-a185-44c2-ba1c-f05b3b420add {{(pid=63372) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1345.935759] env[63372]: DEBUG nova.scheduler.client.report [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Inventory has not changed for provider c1b110ca-a185-44c2-ba1c-f05b3b420add based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 186, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63372) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1346.441073] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.631s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1346.464063] env[63372]: INFO nova.scheduler.client.report [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Deleted allocations for instance 3934ef28-318d-40a6-a7fd-becfbdaaa978 [ 1346.972145] env[63372]: DEBUG oslo_concurrency.lockutils [None req-e671a177-23eb-4280-bf07-c74ec8703ac5 tempest-AttachVolumeNegativeTest-1738950967 tempest-AttachVolumeNegativeTest-1738950967-project-member] Lock "3934ef28-318d-40a6-a7fd-becfbdaaa978" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.212s {{(pid=63372) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}